Cleanup usage of HEAP in mark-compact.

Review URL: http://codereview.chromium.org/6760025

git-svn-id: http://v8.googlecode.com/svn/branches/bleeding_edge@7475 ce2b1a6d-e550-0410-aec6-3dcde31c8c00
This commit is contained in:
karlklose@chromium.org 2011-04-01 11:59:00 +00:00
parent 1e8079fcc9
commit b8d5fd7d32
5 changed files with 211 additions and 203 deletions

View File

@ -223,9 +223,9 @@ void RelocInfo::Visit(Heap* heap) {
if (mode == RelocInfo::EMBEDDED_OBJECT) {
StaticVisitor::VisitPointer(heap, target_object_address());
} else if (RelocInfo::IsCodeTarget(mode)) {
StaticVisitor::VisitCodeTarget(this);
StaticVisitor::VisitCodeTarget(heap, this);
} else if (mode == RelocInfo::GLOBAL_PROPERTY_CELL) {
StaticVisitor::VisitGlobalPropertyCell(this);
StaticVisitor::VisitGlobalPropertyCell(heap, this);
} else if (mode == RelocInfo::EXTERNAL_REFERENCE) {
StaticVisitor::VisitExternalReference(target_reference_address());
#ifdef ENABLE_DEBUGGER_SUPPORT
@ -234,7 +234,7 @@ void RelocInfo::Visit(Heap* heap) {
IsPatchedReturnSequence()) ||
(RelocInfo::IsDebugBreakSlot(mode) &&
IsPatchedDebugBreakSlotSequence()))) {
StaticVisitor::VisitDebugTarget(this);
StaticVisitor::VisitDebugTarget(heap, this);
#endif
} else if (mode == RelocInfo::RUNTIME_ENTRY) {
StaticVisitor::VisitRuntimeEntry(this);

View File

@ -225,9 +225,9 @@ void RelocInfo::Visit(Heap* heap) {
StaticVisitor::VisitPointer(heap, target_object_address());
CPU::FlushICache(pc_, sizeof(Address));
} else if (RelocInfo::IsCodeTarget(mode)) {
StaticVisitor::VisitCodeTarget(this);
StaticVisitor::VisitCodeTarget(heap, this);
} else if (mode == RelocInfo::GLOBAL_PROPERTY_CELL) {
StaticVisitor::VisitGlobalPropertyCell(this);
StaticVisitor::VisitGlobalPropertyCell(heap, this);
} else if (mode == RelocInfo::EXTERNAL_REFERENCE) {
StaticVisitor::VisitExternalReference(target_reference_address());
CPU::FlushICache(pc_, sizeof(Address));
@ -237,7 +237,7 @@ void RelocInfo::Visit(Heap* heap) {
IsPatchedReturnSequence()) ||
(RelocInfo::IsDebugBreakSlot(mode) &&
IsPatchedDebugBreakSlotSequence()))) {
StaticVisitor::VisitDebugTarget(this);
StaticVisitor::VisitDebugTarget(heap, this);
#endif
} else if (mode == RelocInfo::RUNTIME_ENTRY) {
StaticVisitor::VisitRuntimeEntry(this);

View File

@ -86,15 +86,15 @@ void MarkCompactCollector::CollectGarbage() {
GCTracer::Scope gc_scope(tracer_, GCTracer::Scope::MC_COMPACT);
EncodeForwardingAddresses();
heap_->MarkMapPointersAsEncoded(true);
heap()->MarkMapPointersAsEncoded(true);
UpdatePointers();
heap_->MarkMapPointersAsEncoded(false);
heap_->isolate()->pc_to_code_cache()->Flush();
heap()->MarkMapPointersAsEncoded(false);
heap()->isolate()->pc_to_code_cache()->Flush();
RelocateObjects();
} else {
SweepSpaces();
heap_->isolate()->pc_to_code_cache()->Flush();
heap()->isolate()->pc_to_code_cache()->Flush();
}
Finish();
@ -123,7 +123,7 @@ void MarkCompactCollector::Prepare(GCTracer* tracer) {
compact_on_next_gc_ = false;
if (FLAG_never_compact) compacting_collection_ = false;
if (!HEAP->map_space()->MapPointersEncodable())
if (!heap()->map_space()->MapPointersEncodable())
compacting_collection_ = false;
if (FLAG_collect_maps) CreateBackPointers();
#ifdef ENABLE_GDB_JIT_INTERFACE
@ -161,9 +161,9 @@ void MarkCompactCollector::Finish() {
// force lazy re-initialization of it. This must be done after the
// GC, because it relies on the new address of certain old space
// objects (empty string, illegal builtin).
heap_->isolate()->stub_cache()->Clear();
heap()->isolate()->stub_cache()->Clear();
heap_->external_string_table_.CleanUp();
heap()->external_string_table_.CleanUp();
// If we've just compacted old space there's no reason to check the
// fragmentation limit. Just return.
@ -456,7 +456,7 @@ class StaticMarkingVisitor : public StaticVisitorBase {
for (Object** p = start; p < end; p++) MarkObjectByPointer(heap, p);
}
static inline void VisitCodeTarget(RelocInfo* rinfo) {
static inline void VisitCodeTarget(Heap* heap, RelocInfo* rinfo) {
ASSERT(RelocInfo::IsCodeTarget(rinfo->rmode()));
Code* code = Code::GetCodeFromTargetAddress(rinfo->target_address());
if (FLAG_cleanup_ics_at_gc && code->is_inline_cache_stub()) {
@ -464,28 +464,27 @@ class StaticMarkingVisitor : public StaticVisitorBase {
// Please note targets for cleared inline cached do not have to be
// marked since they are contained in HEAP->non_monomorphic_cache().
} else {
code->heap()->mark_compact_collector()->MarkObject(code);
heap->mark_compact_collector()->MarkObject(code);
}
}
static void VisitGlobalPropertyCell(RelocInfo* rinfo) {
static void VisitGlobalPropertyCell(Heap* heap, RelocInfo* rinfo) {
ASSERT(rinfo->rmode() == RelocInfo::GLOBAL_PROPERTY_CELL);
Object* cell = rinfo->target_cell();
Object* old_cell = cell;
VisitPointer(reinterpret_cast<JSGlobalPropertyCell*>(cell)->heap(), &cell);
VisitPointer(heap, &cell);
if (cell != old_cell) {
rinfo->set_target_cell(reinterpret_cast<JSGlobalPropertyCell*>(cell));
}
}
static inline void VisitDebugTarget(RelocInfo* rinfo) {
static inline void VisitDebugTarget(Heap* heap, RelocInfo* rinfo) {
ASSERT((RelocInfo::IsJSReturn(rinfo->rmode()) &&
rinfo->IsPatchedReturnSequence()) ||
(RelocInfo::IsDebugBreakSlot(rinfo->rmode()) &&
rinfo->IsPatchedDebugBreakSlotSequence()));
HeapObject* code = Code::GetCodeFromTargetAddress(rinfo->call_address());
reinterpret_cast<Code*>(code)->heap()->mark_compact_collector()->
MarkObject(code);
heap->mark_compact_collector()->MarkObject(code);
}
// Mark object pointed to by p.
@ -565,8 +564,8 @@ class StaticMarkingVisitor : public StaticVisitorBase {
// flushed.
static const int kCodeAgeThreshold = 5;
inline static bool HasSourceCode(SharedFunctionInfo* info) {
Object* undefined = HEAP->raw_unchecked_undefined_value();
inline static bool HasSourceCode(Heap* heap, SharedFunctionInfo* info) {
Object* undefined = heap->raw_unchecked_undefined_value();
return (info->script() != undefined) &&
(reinterpret_cast<Script*>(info->script())->source() != undefined);
}
@ -582,7 +581,7 @@ class StaticMarkingVisitor : public StaticVisitorBase {
function->GetIsolate()->builtins()->builtin(Builtins::kLazyCompile);
}
inline static bool IsFlushable(JSFunction* function) {
inline static bool IsFlushable(Heap* heap, JSFunction* function) {
SharedFunctionInfo* shared_info = function->unchecked_shared();
// Code is either on stack, in compilation cache or referenced
@ -597,10 +596,10 @@ class StaticMarkingVisitor : public StaticVisitorBase {
return false;
}
return IsFlushable(shared_info);
return IsFlushable(heap, shared_info);
}
inline static bool IsFlushable(SharedFunctionInfo* shared_info) {
inline static bool IsFlushable(Heap* heap, SharedFunctionInfo* shared_info) {
// Code is either on stack, in compilation cache or referenced
// by optimized version of function.
if (shared_info->unchecked_code()->IsMarked()) {
@ -610,7 +609,7 @@ class StaticMarkingVisitor : public StaticVisitorBase {
// The function must be compiled and have the source code available,
// to be able to recompile it in case we need the function again.
if (!(shared_info->is_compiled() && HasSourceCode(shared_info))) {
if (!(shared_info->is_compiled() && HasSourceCode(heap, shared_info))) {
return false;
}
@ -642,7 +641,7 @@ class StaticMarkingVisitor : public StaticVisitorBase {
static bool FlushCodeForFunction(Heap* heap, JSFunction* function) {
if (!IsFlushable(function)) return false;
if (!IsFlushable(heap, function)) return false;
// This function's code looks flushable. But we have to postpone the
// decision until we see all functions that point to the same
@ -719,7 +718,7 @@ class StaticMarkingVisitor : public StaticVisitorBase {
if (shared->IsInobjectSlackTrackingInProgress()) shared->DetachInitialMap();
if (!known_flush_code_candidate) {
known_flush_code_candidate = IsFlushable(shared);
known_flush_code_candidate = IsFlushable(heap, shared);
if (known_flush_code_candidate) {
heap->mark_compact_collector()->code_flusher()->AddCandidate(shared);
}
@ -869,16 +868,16 @@ class MarkingVisitor : public ObjectVisitor {
StaticMarkingVisitor::VisitPointers(heap_, start, end);
}
void VisitCodeTarget(RelocInfo* rinfo) {
StaticMarkingVisitor::VisitCodeTarget(rinfo);
void VisitCodeTarget(Heap* heap, RelocInfo* rinfo) {
StaticMarkingVisitor::VisitCodeTarget(heap, rinfo);
}
void VisitGlobalPropertyCell(RelocInfo* rinfo) {
StaticMarkingVisitor::VisitGlobalPropertyCell(rinfo);
void VisitGlobalPropertyCell(Heap* heap, RelocInfo* rinfo) {
StaticMarkingVisitor::VisitGlobalPropertyCell(heap, rinfo);
}
void VisitDebugTarget(RelocInfo* rinfo) {
StaticMarkingVisitor::VisitDebugTarget(rinfo);
void VisitDebugTarget(Heap* heap, RelocInfo* rinfo) {
StaticMarkingVisitor::VisitDebugTarget(heap, rinfo);
}
private:
@ -926,7 +925,7 @@ class SharedFunctionInfoMarkingVisitor : public ObjectVisitor {
void MarkCompactCollector::PrepareForCodeFlushing() {
ASSERT(heap_ == Isolate::Current()->heap());
ASSERT(heap() == Isolate::Current()->heap());
if (!FLAG_flush_code) {
EnableCodeFlushing(false);
@ -934,8 +933,8 @@ void MarkCompactCollector::PrepareForCodeFlushing() {
}
#ifdef ENABLE_DEBUGGER_SUPPORT
if (heap_->isolate()->debug()->IsLoaded() ||
heap_->isolate()->debug()->has_break_points()) {
if (heap()->isolate()->debug()->IsLoaded() ||
heap()->isolate()->debug()->has_break_points()) {
EnableCodeFlushing(false);
return;
}
@ -944,10 +943,10 @@ void MarkCompactCollector::PrepareForCodeFlushing() {
// Ensure that empty descriptor array is marked. Method MarkDescriptorArray
// relies on it being marked before any other descriptor array.
MarkObject(heap_->raw_unchecked_empty_descriptor_array());
MarkObject(heap()->raw_unchecked_empty_descriptor_array());
// Make sure we are not referencing the code from the stack.
ASSERT(this == heap_->mark_compact_collector());
ASSERT(this == heap()->mark_compact_collector());
for (StackFrameIterator it; !it.done(); it.Advance()) {
MarkObject(it.frame()->unchecked_code());
}
@ -955,12 +954,12 @@ void MarkCompactCollector::PrepareForCodeFlushing() {
// Iterate the archived stacks in all threads to check if
// the code is referenced.
CodeMarkingVisitor code_marking_visitor(this);
heap_->isolate()->thread_manager()->IterateArchivedThreads(
heap()->isolate()->thread_manager()->IterateArchivedThreads(
&code_marking_visitor);
SharedFunctionInfoMarkingVisitor visitor(this);
heap_->isolate()->compilation_cache()->IterateFunctions(&visitor);
heap_->isolate()->handle_scope_implementer()->Iterate(&visitor);
heap()->isolate()->compilation_cache()->IterateFunctions(&visitor);
heap()->isolate()->handle_scope_implementer()->Iterate(&visitor);
ProcessMarkingStack();
}
@ -1008,7 +1007,8 @@ class RootMarkingVisitor : public ObjectVisitor {
// Helper class for pruning the symbol table.
class SymbolTableCleaner : public ObjectVisitor {
public:
SymbolTableCleaner() : pointers_removed_(0) { }
explicit SymbolTableCleaner(Heap* heap)
: heap_(heap), pointers_removed_(0) { }
virtual void VisitPointers(Object** start, Object** end) {
// Visit all HeapObject pointers in [start, end).
@ -1020,10 +1020,10 @@ class SymbolTableCleaner : public ObjectVisitor {
// Since no objects have yet been moved we can safely access the map of
// the object.
if ((*p)->IsExternalString()) {
HEAP->FinalizeExternalString(String::cast(*p));
heap_->FinalizeExternalString(String::cast(*p));
}
// Set the entry to null_value (as deleted).
*p = HEAP->raw_unchecked_null_value();
*p = heap_->raw_unchecked_null_value();
pointers_removed_++;
}
}
@ -1033,6 +1033,7 @@ class SymbolTableCleaner : public ObjectVisitor {
return pointers_removed_;
}
private:
Heap* heap_;
int pointers_removed_;
};
@ -1058,7 +1059,7 @@ void MarkCompactCollector::MarkUnmarkedObject(HeapObject* object) {
if (object->IsMap()) {
Map* map = Map::cast(object);
if (FLAG_cleanup_caches_in_maps_at_gc) {
map->ClearCodeCache(heap_);
map->ClearCodeCache(heap());
}
SetMark(map);
if (FLAG_collect_maps &&
@ -1129,7 +1130,7 @@ void MarkCompactCollector::MarkDescriptorArray(
void MarkCompactCollector::CreateBackPointers() {
HeapObjectIterator iterator(HEAP->map_space());
HeapObjectIterator iterator(heap()->map_space());
for (HeapObject* next_object = iterator.next();
next_object != NULL; next_object = iterator.next()) {
if (next_object->IsMap()) { // Could also be ByteArray on free list.
@ -1138,7 +1139,7 @@ void MarkCompactCollector::CreateBackPointers() {
map->instance_type() <= JS_FUNCTION_TYPE) {
map->CreateBackPointers();
} else {
ASSERT(map->instance_descriptors() == HEAP->empty_descriptor_array());
ASSERT(map->instance_descriptors() == heap()->empty_descriptor_array());
}
}
}
@ -1186,11 +1187,11 @@ bool MarkCompactCollector::IsUnmarkedHeapObject(Object** p) {
void MarkCompactCollector::MarkSymbolTable() {
SymbolTable* symbol_table = heap_->raw_unchecked_symbol_table();
SymbolTable* symbol_table = heap()->raw_unchecked_symbol_table();
// Mark the symbol table itself.
SetMark(symbol_table);
// Explicitly mark the prefix.
MarkingVisitor marker(heap_);
MarkingVisitor marker(heap());
symbol_table->IteratePrefix(&marker);
ProcessMarkingStack();
}
@ -1199,7 +1200,7 @@ void MarkCompactCollector::MarkSymbolTable() {
void MarkCompactCollector::MarkRoots(RootMarkingVisitor* visitor) {
// Mark the heap roots including global variables, stack variables,
// etc., and all objects reachable from them.
HEAP->IterateStrongRoots(visitor, VISIT_ONLY_STRONG);
heap()->IterateStrongRoots(visitor, VISIT_ONLY_STRONG);
// Handle the symbol table specially.
MarkSymbolTable();
@ -1214,7 +1215,7 @@ void MarkCompactCollector::MarkRoots(RootMarkingVisitor* visitor) {
void MarkCompactCollector::MarkObjectGroups() {
List<ObjectGroup*>* object_groups =
heap_->isolate()->global_handles()->object_groups();
heap()->isolate()->global_handles()->object_groups();
for (int i = 0; i < object_groups->length(); i++) {
ObjectGroup* entry = object_groups->at(i);
@ -1250,7 +1251,7 @@ void MarkCompactCollector::MarkObjectGroups() {
void MarkCompactCollector::MarkImplicitRefGroups() {
List<ImplicitRefGroup*>* ref_groups =
heap_->isolate()->global_handles()->implicit_ref_groups();
heap()->isolate()->global_handles()->implicit_ref_groups();
for (int i = 0; i < ref_groups->length(); i++) {
ImplicitRefGroup* entry = ref_groups->at(i);
@ -1283,7 +1284,7 @@ void MarkCompactCollector::EmptyMarkingStack() {
while (!marking_stack_.is_empty()) {
HeapObject* object = marking_stack_.Pop();
ASSERT(object->IsHeapObject());
ASSERT(heap_->Contains(object));
ASSERT(heap()->Contains(object));
ASSERT(object->IsMarked());
ASSERT(!object->IsOverflowed());
@ -1307,32 +1308,32 @@ void MarkCompactCollector::EmptyMarkingStack() {
void MarkCompactCollector::RefillMarkingStack() {
ASSERT(marking_stack_.overflowed());
SemiSpaceIterator new_it(HEAP->new_space(), &OverflowObjectSize);
SemiSpaceIterator new_it(heap()->new_space(), &OverflowObjectSize);
OverflowedObjectsScanner::ScanOverflowedObjects(this, &new_it);
if (marking_stack_.is_full()) return;
HeapObjectIterator old_pointer_it(HEAP->old_pointer_space(),
HeapObjectIterator old_pointer_it(heap()->old_pointer_space(),
&OverflowObjectSize);
OverflowedObjectsScanner::ScanOverflowedObjects(this, &old_pointer_it);
if (marking_stack_.is_full()) return;
HeapObjectIterator old_data_it(HEAP->old_data_space(), &OverflowObjectSize);
HeapObjectIterator old_data_it(heap()->old_data_space(), &OverflowObjectSize);
OverflowedObjectsScanner::ScanOverflowedObjects(this, &old_data_it);
if (marking_stack_.is_full()) return;
HeapObjectIterator code_it(HEAP->code_space(), &OverflowObjectSize);
HeapObjectIterator code_it(heap()->code_space(), &OverflowObjectSize);
OverflowedObjectsScanner::ScanOverflowedObjects(this, &code_it);
if (marking_stack_.is_full()) return;
HeapObjectIterator map_it(HEAP->map_space(), &OverflowObjectSize);
HeapObjectIterator map_it(heap()->map_space(), &OverflowObjectSize);
OverflowedObjectsScanner::ScanOverflowedObjects(this, &map_it);
if (marking_stack_.is_full()) return;
HeapObjectIterator cell_it(HEAP->cell_space(), &OverflowObjectSize);
HeapObjectIterator cell_it(heap()->cell_space(), &OverflowObjectSize);
OverflowedObjectsScanner::ScanOverflowedObjects(this, &cell_it);
if (marking_stack_.is_full()) return;
LargeObjectIterator lo_it(HEAP->lo_space(), &OverflowObjectSize);
LargeObjectIterator lo_it(heap()->lo_space(), &OverflowObjectSize);
OverflowedObjectsScanner::ScanOverflowedObjects(this, &lo_it);
if (marking_stack_.is_full()) return;
@ -1370,7 +1371,7 @@ void MarkCompactCollector::MarkLiveObjects() {
// The recursive GC marker detects when it is nearing stack overflow,
// and switches to a different marking system. JS interrupts interfere
// with the C stack limit check.
PostponeInterruptsScope postpone(heap_->isolate());
PostponeInterruptsScope postpone(heap()->isolate());
#ifdef DEBUG
ASSERT(state_ == PREPARE_GC);
@ -1378,14 +1379,14 @@ void MarkCompactCollector::MarkLiveObjects() {
#endif
// The to space contains live objects, the from space is used as a marking
// stack.
marking_stack_.Initialize(heap_->new_space()->FromSpaceLow(),
heap_->new_space()->FromSpaceHigh());
marking_stack_.Initialize(heap()->new_space()->FromSpaceLow(),
heap()->new_space()->FromSpaceHigh());
ASSERT(!marking_stack_.overflowed());
PrepareForCodeFlushing();
RootMarkingVisitor root_visitor(heap_);
RootMarkingVisitor root_visitor(heap());
MarkRoots(&root_visitor);
// The objects reachable from the roots are marked, yet unreachable
@ -1399,10 +1400,10 @@ void MarkCompactCollector::MarkLiveObjects() {
//
// First we identify nonlive weak handles and mark them as pending
// destruction.
heap_->isolate()->global_handles()->IdentifyWeakHandles(
heap()->isolate()->global_handles()->IdentifyWeakHandles(
&IsUnmarkedHeapObject);
// Then we mark the objects and process the transitive closure.
heap_->isolate()->global_handles()->IterateWeakRoots(&root_visitor);
heap()->isolate()->global_handles()->IterateWeakRoots(&root_visitor);
while (marking_stack_.overflowed()) {
RefillMarkingStack();
EmptyMarkingStack();
@ -1415,20 +1416,20 @@ void MarkCompactCollector::MarkLiveObjects() {
// Prune the symbol table removing all symbols only pointed to by the
// symbol table. Cannot use symbol_table() here because the symbol
// table is marked.
SymbolTable* symbol_table = heap_->raw_unchecked_symbol_table();
SymbolTableCleaner v;
SymbolTable* symbol_table = heap()->raw_unchecked_symbol_table();
SymbolTableCleaner v(heap());
symbol_table->IterateElements(&v);
symbol_table->ElementsRemoved(v.PointersRemoved());
heap_->external_string_table_.Iterate(&v);
heap_->external_string_table_.CleanUp();
heap()->external_string_table_.Iterate(&v);
heap()->external_string_table_.CleanUp();
// Process the weak references.
MarkCompactWeakObjectRetainer mark_compact_object_retainer;
heap_->ProcessWeakReferences(&mark_compact_object_retainer);
heap()->ProcessWeakReferences(&mark_compact_object_retainer);
// Remove object groups after marking phase.
heap_->isolate()->global_handles()->RemoveObjectGroups();
heap_->isolate()->global_handles()->RemoveImplicitRefGroups();
heap()->isolate()->global_handles()->RemoveObjectGroups();
heap()->isolate()->global_handles()->RemoveImplicitRefGroups();
// Flush code from collected candidates.
if (is_code_flushing_enabled()) {
@ -1436,28 +1437,28 @@ void MarkCompactCollector::MarkLiveObjects() {
}
// Clean up dead objects from the runtime profiler.
heap_->isolate()->runtime_profiler()->RemoveDeadSamples();
heap()->isolate()->runtime_profiler()->RemoveDeadSamples();
}
#ifdef DEBUG
void MarkCompactCollector::UpdateLiveObjectCount(HeapObject* obj) {
live_bytes_ += obj->Size();
if (HEAP->new_space()->Contains(obj)) {
if (heap()->new_space()->Contains(obj)) {
live_young_objects_size_ += obj->Size();
} else if (HEAP->map_space()->Contains(obj)) {
} else if (heap()->map_space()->Contains(obj)) {
ASSERT(obj->IsMap());
live_map_objects_size_ += obj->Size();
} else if (HEAP->cell_space()->Contains(obj)) {
} else if (heap()->cell_space()->Contains(obj)) {
ASSERT(obj->IsJSGlobalPropertyCell());
live_cell_objects_size_ += obj->Size();
} else if (HEAP->old_pointer_space()->Contains(obj)) {
} else if (heap()->old_pointer_space()->Contains(obj)) {
live_old_pointer_objects_size_ += obj->Size();
} else if (HEAP->old_data_space()->Contains(obj)) {
} else if (heap()->old_data_space()->Contains(obj)) {
live_old_data_objects_size_ += obj->Size();
} else if (HEAP->code_space()->Contains(obj)) {
} else if (heap()->code_space()->Contains(obj)) {
live_code_objects_size_ += obj->Size();
} else if (HEAP->lo_space()->Contains(obj)) {
} else if (heap()->lo_space()->Contains(obj)) {
live_lo_objects_size_ += obj->Size();
} else {
UNREACHABLE();
@ -1473,7 +1474,7 @@ void MarkCompactCollector::SweepLargeObjectSpace() {
compacting_collection_ ? ENCODE_FORWARDING_ADDRESSES : SWEEP_SPACES;
#endif
// Deallocate unmarked objects and clear marked bits for marked objects.
HEAP->lo_space()->FreeUnmarkedObjects();
heap()->lo_space()->FreeUnmarkedObjects();
}
@ -1486,7 +1487,7 @@ bool MarkCompactCollector::SafeIsMap(HeapObject* object) {
void MarkCompactCollector::ClearNonLiveTransitions() {
HeapObjectIterator map_iterator(HEAP->map_space(), &SizeOfMarkedObject);
HeapObjectIterator map_iterator(heap() ->map_space(), &SizeOfMarkedObject);
// Iterate over the map space, setting map transitions that go from
// a marked map to an unmarked map to null transitions. At the same time,
// set all the prototype fields of maps back to their original value,
@ -1536,7 +1537,7 @@ void MarkCompactCollector::ClearNonLiveTransitions() {
// This test will always be false on the first iteration.
if (on_dead_path && current->IsMarked()) {
on_dead_path = false;
current->ClearNonLiveTransitions(heap_, real_prototype);
current->ClearNonLiveTransitions(heap(), real_prototype);
}
*HeapObject::RawField(current, Map::kPrototypeOffset) =
real_prototype;
@ -1771,8 +1772,8 @@ void MarkCompactCollector::EncodeForwardingAddressesInNewSpace() {
EncodeForwardingAddressInNewSpace,
IgnoreNonLiveObject>(
this,
heap_->new_space()->bottom(),
heap_->new_space()->top(),
heap()->new_space()->bottom(),
heap()->new_space()->top(),
&ignored);
}
@ -2194,24 +2195,24 @@ void MarkCompactCollector::EncodeForwardingAddresses() {
// Objects in the active semispace of the young generation may be
// relocated to the inactive semispace (if not promoted). Set the
// relocation info to the beginning of the inactive semispace.
heap_->new_space()->MCResetRelocationInfo();
heap()->new_space()->MCResetRelocationInfo();
// Compute the forwarding pointers in each space.
EncodeForwardingAddressesInPagedSpace<MCAllocateFromOldPointerSpace,
ReportDeleteIfNeeded>(
heap_->old_pointer_space());
heap()->old_pointer_space());
EncodeForwardingAddressesInPagedSpace<MCAllocateFromOldDataSpace,
IgnoreNonLiveObject>(
heap_->old_data_space());
heap()->old_data_space());
EncodeForwardingAddressesInPagedSpace<MCAllocateFromCodeSpace,
ReportDeleteIfNeeded>(
heap_->code_space());
heap()->code_space());
EncodeForwardingAddressesInPagedSpace<MCAllocateFromCellSpace,
IgnoreNonLiveObject>(
heap_->cell_space());
heap()->cell_space());
// Compute new space next to last after the old and code spaces have been
@ -2223,25 +2224,26 @@ void MarkCompactCollector::EncodeForwardingAddresses() {
// non-live map pointers to get the sizes of non-live objects.
EncodeForwardingAddressesInPagedSpace<MCAllocateFromMapSpace,
IgnoreNonLiveObject>(
heap_->map_space());
heap()->map_space());
// Write relocation info to the top page, so we can use it later. This is
// done after promoting objects from the new space so we get the correct
// allocation top.
heap_->old_pointer_space()->MCWriteRelocationInfoToPage();
heap_->old_data_space()->MCWriteRelocationInfoToPage();
heap_->code_space()->MCWriteRelocationInfoToPage();
heap_->map_space()->MCWriteRelocationInfoToPage();
heap_->cell_space()->MCWriteRelocationInfoToPage();
heap()->old_pointer_space()->MCWriteRelocationInfoToPage();
heap()->old_data_space()->MCWriteRelocationInfoToPage();
heap()->code_space()->MCWriteRelocationInfoToPage();
heap()->map_space()->MCWriteRelocationInfoToPage();
heap()->cell_space()->MCWriteRelocationInfoToPage();
}
class MapIterator : public HeapObjectIterator {
public:
MapIterator() : HeapObjectIterator(HEAP->map_space(), &SizeCallback) { }
explicit MapIterator(Heap* heap)
: HeapObjectIterator(heap->map_space(), &SizeCallback) { }
explicit MapIterator(Address start)
: HeapObjectIterator(HEAP->map_space(), start, &SizeCallback) { }
MapIterator(Heap* heap, Address start)
: HeapObjectIterator(heap->map_space(), start, &SizeCallback) { }
private:
static int SizeCallback(HeapObject* unused) {
@ -2257,7 +2259,8 @@ class MapCompact {
: heap_(heap),
live_maps_(live_maps),
to_evacuate_start_(heap->map_space()->TopAfterCompaction(live_maps)),
map_to_evacuate_it_(to_evacuate_start_),
vacant_map_it_(heap),
map_to_evacuate_it_(heap, to_evacuate_start_),
first_map_to_evacuate_(
reinterpret_cast<Map*>(HeapObject::FromAddress(to_evacuate_start_))) {
}
@ -2278,36 +2281,41 @@ class MapCompact {
void UpdateMapPointersInRoots() {
MapUpdatingVisitor map_updating_visitor;
heap_->IterateRoots(&map_updating_visitor, VISIT_ONLY_STRONG);
heap_->isolate()->global_handles()->IterateWeakRoots(&map_updating_visitor);
heap()->IterateRoots(&map_updating_visitor, VISIT_ONLY_STRONG);
heap()->isolate()->global_handles()->IterateWeakRoots(
&map_updating_visitor);
LiveObjectList::IterateElements(&map_updating_visitor);
}
void UpdateMapPointersInPagedSpace(PagedSpace* space) {
ASSERT(space != heap_->map_space());
ASSERT(space != heap()->map_space());
PageIterator it(space, PageIterator::PAGES_IN_USE);
while (it.has_next()) {
Page* p = it.next();
UpdateMapPointersInRange(heap_, p->ObjectAreaStart(), p->AllocationTop());
UpdateMapPointersInRange(heap(),
p->ObjectAreaStart(),
p->AllocationTop());
}
}
void UpdateMapPointersInNewSpace() {
NewSpace* space = heap_->new_space();
UpdateMapPointersInRange(heap_, space->bottom(), space->top());
NewSpace* space = heap()->new_space();
UpdateMapPointersInRange(heap(), space->bottom(), space->top());
}
void UpdateMapPointersInLargeObjectSpace() {
LargeObjectIterator it(heap_->lo_space());
LargeObjectIterator it(heap()->lo_space());
for (HeapObject* obj = it.next(); obj != NULL; obj = it.next())
UpdateMapPointersInObject(heap_, obj);
UpdateMapPointersInObject(heap(), obj);
}
void Finish() {
heap_->map_space()->FinishCompaction(to_evacuate_start_, live_maps_);
heap()->map_space()->FinishCompaction(to_evacuate_start_, live_maps_);
}
inline Heap* heap() const { return heap_; }
private:
Heap* heap_;
int live_maps_;
@ -2457,26 +2465,26 @@ void MarkCompactCollector::SweepSpaces() {
// the map space last because freeing non-live maps overwrites them and
// the other spaces rely on possibly non-live maps to get the sizes for
// non-live objects.
SweepSpace(heap_, heap_->old_pointer_space());
SweepSpace(heap_, heap_->old_data_space());
SweepSpace(heap_, heap_->code_space());
SweepSpace(heap_, heap_->cell_space());
SweepSpace(heap(), heap()->old_pointer_space());
SweepSpace(heap(), heap()->old_data_space());
SweepSpace(heap(), heap()->code_space());
SweepSpace(heap(), heap()->cell_space());
{ GCTracer::Scope gc_scope(tracer_, GCTracer::Scope::MC_SWEEP_NEWSPACE);
SweepNewSpace(heap_, heap_->new_space());
SweepNewSpace(heap(), heap()->new_space());
}
SweepSpace(heap_, heap_->map_space());
SweepSpace(heap(), heap()->map_space());
heap_->IterateDirtyRegions(heap_->map_space(),
&heap_->IteratePointersInDirtyMapsRegion,
heap()->IterateDirtyRegions(heap()->map_space(),
&heap()->IteratePointersInDirtyMapsRegion,
&UpdatePointerToNewGen,
heap_->WATERMARK_SHOULD_BE_VALID);
heap()->WATERMARK_SHOULD_BE_VALID);
intptr_t live_maps_size = heap_->map_space()->Size();
intptr_t live_maps_size = heap()->map_space()->Size();
int live_maps = static_cast<int>(live_maps_size / Map::kSize);
ASSERT(live_map_objects_size_ == live_maps_size);
if (heap_->map_space()->NeedsCompaction(live_maps)) {
MapCompact map_compact(heap_, live_maps);
if (heap()->map_space()->NeedsCompaction(live_maps)) {
MapCompact map_compact(heap(), live_maps);
map_compact.CompactMaps();
map_compact.UpdateMapPointersInRoots();
@ -2484,7 +2492,7 @@ void MarkCompactCollector::SweepSpaces() {
PagedSpaces spaces;
for (PagedSpace* space = spaces.next();
space != NULL; space = spaces.next()) {
if (space == heap_->map_space()) continue;
if (space == heap()->map_space()) continue;
map_compact.UpdateMapPointersInPagedSpace(space);
}
map_compact.UpdateMapPointersInNewSpace();
@ -2580,6 +2588,8 @@ class UpdatingVisitor: public ObjectVisitor {
reinterpret_cast<Code*>(target)->instruction_start());
}
inline Heap* heap() const { return heap_; }
private:
void UpdatePointer(Object** p) {
if (!(*p)->IsHeapObject()) return;
@ -2587,27 +2597,27 @@ class UpdatingVisitor: public ObjectVisitor {
HeapObject* obj = HeapObject::cast(*p);
Address old_addr = obj->address();
Address new_addr;
ASSERT(!heap_->InFromSpace(obj));
ASSERT(!heap()->InFromSpace(obj));
if (heap_->new_space()->Contains(obj)) {
if (heap()->new_space()->Contains(obj)) {
Address forwarding_pointer_addr =
heap_->new_space()->FromSpaceLow() +
heap_->new_space()->ToSpaceOffsetForAddress(old_addr);
heap()->new_space()->FromSpaceLow() +
heap()->new_space()->ToSpaceOffsetForAddress(old_addr);
new_addr = Memory::Address_at(forwarding_pointer_addr);
#ifdef DEBUG
ASSERT(heap_->old_pointer_space()->Contains(new_addr) ||
heap_->old_data_space()->Contains(new_addr) ||
heap_->new_space()->FromSpaceContains(new_addr) ||
heap_->lo_space()->Contains(HeapObject::FromAddress(new_addr)));
ASSERT(heap()->old_pointer_space()->Contains(new_addr) ||
heap()->old_data_space()->Contains(new_addr) ||
heap()->new_space()->FromSpaceContains(new_addr) ||
heap()->lo_space()->Contains(HeapObject::FromAddress(new_addr)));
if (heap_->new_space()->FromSpaceContains(new_addr)) {
ASSERT(heap_->new_space()->FromSpaceOffsetForAddress(new_addr) <=
heap_->new_space()->ToSpaceOffsetForAddress(old_addr));
if (heap()->new_space()->FromSpaceContains(new_addr)) {
ASSERT(heap()->new_space()->FromSpaceOffsetForAddress(new_addr) <=
heap()->new_space()->ToSpaceOffsetForAddress(old_addr));
}
#endif
} else if (heap_->lo_space()->Contains(obj)) {
} else if (heap()->lo_space()->Contains(obj)) {
// Don't move objects in the large object space.
return;
@ -2646,34 +2656,34 @@ void MarkCompactCollector::UpdatePointers() {
ASSERT(state_ == ENCODE_FORWARDING_ADDRESSES);
state_ = UPDATE_POINTERS;
#endif
UpdatingVisitor updating_visitor(heap_);
heap_->isolate()->runtime_profiler()->UpdateSamplesAfterCompact(
UpdatingVisitor updating_visitor(heap());
heap()->isolate()->runtime_profiler()->UpdateSamplesAfterCompact(
&updating_visitor);
heap_->IterateRoots(&updating_visitor, VISIT_ONLY_STRONG);
heap_->isolate()->global_handles()->IterateWeakRoots(&updating_visitor);
heap()->IterateRoots(&updating_visitor, VISIT_ONLY_STRONG);
heap()->isolate()->global_handles()->IterateWeakRoots(&updating_visitor);
// Update the pointer to the head of the weak list of global contexts.
updating_visitor.VisitPointer(&heap_->global_contexts_list_);
updating_visitor.VisitPointer(&heap()->global_contexts_list_);
LiveObjectList::IterateElements(&updating_visitor);
int live_maps_size = IterateLiveObjects(
heap_->map_space(), &MarkCompactCollector::UpdatePointersInOldObject);
heap()->map_space(), &MarkCompactCollector::UpdatePointersInOldObject);
int live_pointer_olds_size = IterateLiveObjects(
heap_->old_pointer_space(),
heap()->old_pointer_space(),
&MarkCompactCollector::UpdatePointersInOldObject);
int live_data_olds_size = IterateLiveObjects(
heap_->old_data_space(),
heap()->old_data_space(),
&MarkCompactCollector::UpdatePointersInOldObject);
int live_codes_size = IterateLiveObjects(
heap_->code_space(), &MarkCompactCollector::UpdatePointersInOldObject);
heap()->code_space(), &MarkCompactCollector::UpdatePointersInOldObject);
int live_cells_size = IterateLiveObjects(
heap_->cell_space(), &MarkCompactCollector::UpdatePointersInOldObject);
heap()->cell_space(), &MarkCompactCollector::UpdatePointersInOldObject);
int live_news_size = IterateLiveObjects(
heap_->new_space(), &MarkCompactCollector::UpdatePointersInNewObject);
heap()->new_space(), &MarkCompactCollector::UpdatePointersInNewObject);
// Large objects do not move, the map word can be updated directly.
LargeObjectIterator it(heap_->lo_space());
LargeObjectIterator it(heap()->lo_space());
for (HeapObject* obj = it.next(); obj != NULL; obj = it.next()) {
UpdatePointersInNewObject(obj);
}
@ -2700,8 +2710,8 @@ int MarkCompactCollector::UpdatePointersInNewObject(HeapObject* obj) {
Address forwarded = GetForwardingAddressInOldSpace(old_map);
ASSERT(heap_->map_space()->Contains(old_map));
ASSERT(heap_->map_space()->Contains(forwarded));
ASSERT(heap()->map_space()->Contains(old_map));
ASSERT(heap()->map_space()->Contains(forwarded));
#ifdef DEBUG
if (FLAG_gc_verbose) {
PrintF("update %p : %p -> %p\n", obj->address(), old_map->address(),
@ -2716,7 +2726,7 @@ int MarkCompactCollector::UpdatePointersInNewObject(HeapObject* obj) {
int obj_size = obj->SizeFromMap(old_map);
// Update pointers in the object body.
UpdatingVisitor updating_visitor(heap_);
UpdatingVisitor updating_visitor(heap());
obj->IterateBody(old_map->instance_type(), obj_size, &updating_visitor);
return obj_size;
}
@ -2725,8 +2735,8 @@ int MarkCompactCollector::UpdatePointersInNewObject(HeapObject* obj) {
int MarkCompactCollector::UpdatePointersInOldObject(HeapObject* obj) {
// Decode the map pointer.
MapWord encoding = obj->map_word();
Address map_addr = encoding.DecodeMapAddress(heap_->map_space());
ASSERT(heap_->map_space()->Contains(HeapObject::FromAddress(map_addr)));
Address map_addr = encoding.DecodeMapAddress(heap()->map_space());
ASSERT(heap()->map_space()->Contains(HeapObject::FromAddress(map_addr)));
// At this point, the first word of map_addr is also encoded, cannot
// cast it to Map* using Map::cast.
@ -2747,7 +2757,7 @@ int MarkCompactCollector::UpdatePointersInOldObject(HeapObject* obj) {
#endif
// Update pointers in the object body.
UpdatingVisitor updating_visitor(heap_);
UpdatingVisitor updating_visitor(heap());
obj->IterateBody(type, obj_size, &updating_visitor);
return obj_size;
}
@ -2804,18 +2814,18 @@ void MarkCompactCollector::RelocateObjects() {
// Relocates objects, always relocate map objects first. Relocating
// objects in other space relies on map objects to get object size.
int live_maps_size = IterateLiveObjects(
heap_->map_space(), &MarkCompactCollector::RelocateMapObject);
heap()->map_space(), &MarkCompactCollector::RelocateMapObject);
int live_pointer_olds_size = IterateLiveObjects(
heap_->old_pointer_space(),
heap()->old_pointer_space(),
&MarkCompactCollector::RelocateOldPointerObject);
int live_data_olds_size = IterateLiveObjects(
heap_->old_data_space(), &MarkCompactCollector::RelocateOldDataObject);
heap()->old_data_space(), &MarkCompactCollector::RelocateOldDataObject);
int live_codes_size = IterateLiveObjects(
heap_->code_space(), &MarkCompactCollector::RelocateCodeObject);
heap()->code_space(), &MarkCompactCollector::RelocateCodeObject);
int live_cells_size = IterateLiveObjects(
heap_->cell_space(), &MarkCompactCollector::RelocateCellObject);
heap()->cell_space(), &MarkCompactCollector::RelocateCellObject);
int live_news_size = IterateLiveObjects(
heap_->new_space(), &MarkCompactCollector::RelocateNewObject);
heap()->new_space(), &MarkCompactCollector::RelocateNewObject);
USE(live_maps_size);
USE(live_pointer_olds_size);
@ -2831,28 +2841,28 @@ void MarkCompactCollector::RelocateObjects() {
ASSERT(live_news_size == live_young_objects_size_);
// Flip from and to spaces
heap_->new_space()->Flip();
heap()->new_space()->Flip();
heap_->new_space()->MCCommitRelocationInfo();
heap()->new_space()->MCCommitRelocationInfo();
// Set age_mark to bottom in to space
Address mark = heap_->new_space()->bottom();
heap_->new_space()->set_age_mark(mark);
Address mark = heap()->new_space()->bottom();
heap()->new_space()->set_age_mark(mark);
PagedSpaces spaces;
for (PagedSpace* space = spaces.next(); space != NULL; space = spaces.next())
space->MCCommitRelocationInfo();
heap_->CheckNewSpaceExpansionCriteria();
heap_->IncrementYoungSurvivorsCounter(live_news_size);
heap()->CheckNewSpaceExpansionCriteria();
heap()->IncrementYoungSurvivorsCounter(live_news_size);
}
int MarkCompactCollector::RelocateMapObject(HeapObject* obj) {
// Recover map pointer.
MapWord encoding = obj->map_word();
Address map_addr = encoding.DecodeMapAddress(heap_->map_space());
ASSERT(heap_->map_space()->Contains(HeapObject::FromAddress(map_addr)));
Address map_addr = encoding.DecodeMapAddress(heap()->map_space());
ASSERT(heap()->map_space()->Contains(HeapObject::FromAddress(map_addr)));
// Get forwarding address before resetting map pointer
Address new_addr = GetForwardingAddressInOldSpace(obj);
@ -2865,7 +2875,7 @@ int MarkCompactCollector::RelocateMapObject(HeapObject* obj) {
if (new_addr != old_addr) {
// Move contents.
heap_->MoveBlockToOldSpaceAndUpdateRegionMarks(new_addr,
heap()->MoveBlockToOldSpaceAndUpdateRegionMarks(new_addr,
old_addr,
Map::kSize);
}
@ -2911,8 +2921,8 @@ int MarkCompactCollector::RelocateOldNonCodeObject(HeapObject* obj,
PagedSpace* space) {
// Recover map pointer.
MapWord encoding = obj->map_word();
Address map_addr = encoding.DecodeMapAddress(heap_->map_space());
ASSERT(heap_->map_space()->Contains(map_addr));
Address map_addr = encoding.DecodeMapAddress(heap()->map_space());
ASSERT(heap()->map_space()->Contains(map_addr));
// Get forwarding address before resetting map pointer.
Address new_addr = GetForwardingAddressInOldSpace(obj);
@ -2924,10 +2934,10 @@ int MarkCompactCollector::RelocateOldNonCodeObject(HeapObject* obj,
if (new_addr != old_addr) {
// Move contents.
if (space == heap_->old_data_space()) {
heap_->MoveBlock(new_addr, old_addr, obj_size);
if (space == heap()->old_data_space()) {
heap()->MoveBlock(new_addr, old_addr, obj_size);
} else {
heap_->MoveBlockToOldSpaceAndUpdateRegionMarks(new_addr,
heap()->MoveBlockToOldSpaceAndUpdateRegionMarks(new_addr,
old_addr,
obj_size);
}
@ -2937,47 +2947,47 @@ int MarkCompactCollector::RelocateOldNonCodeObject(HeapObject* obj,
HeapObject* copied_to = HeapObject::FromAddress(new_addr);
if (copied_to->IsSharedFunctionInfo()) {
PROFILE(heap_->isolate(),
PROFILE(heap()->isolate(),
SharedFunctionInfoMoveEvent(old_addr, new_addr));
}
HEAP_PROFILE(heap_, ObjectMoveEvent(old_addr, new_addr));
HEAP_PROFILE(heap(), ObjectMoveEvent(old_addr, new_addr));
return obj_size;
}
int MarkCompactCollector::RelocateOldPointerObject(HeapObject* obj) {
return RelocateOldNonCodeObject(obj, heap_->old_pointer_space());
return RelocateOldNonCodeObject(obj, heap()->old_pointer_space());
}
int MarkCompactCollector::RelocateOldDataObject(HeapObject* obj) {
return RelocateOldNonCodeObject(obj, heap_->old_data_space());
return RelocateOldNonCodeObject(obj, heap()->old_data_space());
}
int MarkCompactCollector::RelocateCellObject(HeapObject* obj) {
return RelocateOldNonCodeObject(obj, heap_->cell_space());
return RelocateOldNonCodeObject(obj, heap()->cell_space());
}
int MarkCompactCollector::RelocateCodeObject(HeapObject* obj) {
// Recover map pointer.
MapWord encoding = obj->map_word();
Address map_addr = encoding.DecodeMapAddress(heap_->map_space());
ASSERT(heap_->map_space()->Contains(HeapObject::FromAddress(map_addr)));
Address map_addr = encoding.DecodeMapAddress(heap()->map_space());
ASSERT(heap()->map_space()->Contains(HeapObject::FromAddress(map_addr)));
// Get forwarding address before resetting map pointer
Address new_addr = GetForwardingAddressInOldSpace(obj);
// Reset the map pointer.
int obj_size = RestoreMap(obj, heap_->code_space(), new_addr, map_addr);
int obj_size = RestoreMap(obj, heap()->code_space(), new_addr, map_addr);
Address old_addr = obj->address();
if (new_addr != old_addr) {
// Move contents.
heap_->MoveBlock(new_addr, old_addr, obj_size);
heap()->MoveBlock(new_addr, old_addr, obj_size);
}
HeapObject* copied_to = HeapObject::FromAddress(new_addr);
@ -2985,9 +2995,9 @@ int MarkCompactCollector::RelocateCodeObject(HeapObject* obj) {
// May also update inline cache target.
Code::cast(copied_to)->Relocate(new_addr - old_addr);
// Notify the logger that compiled code has moved.
PROFILE(heap_->isolate(), CodeMoveEvent(old_addr, new_addr));
PROFILE(heap()->isolate(), CodeMoveEvent(old_addr, new_addr));
}
HEAP_PROFILE(heap_, ObjectMoveEvent(old_addr, new_addr));
HEAP_PROFILE(heap(), ObjectMoveEvent(old_addr, new_addr));
return obj_size;
}
@ -2998,26 +3008,26 @@ int MarkCompactCollector::RelocateNewObject(HeapObject* obj) {
// Get forwarding address
Address old_addr = obj->address();
int offset = heap_->new_space()->ToSpaceOffsetForAddress(old_addr);
int offset = heap()->new_space()->ToSpaceOffsetForAddress(old_addr);
Address new_addr =
Memory::Address_at(heap_->new_space()->FromSpaceLow() + offset);
Memory::Address_at(heap()->new_space()->FromSpaceLow() + offset);
#ifdef DEBUG
if (heap_->new_space()->FromSpaceContains(new_addr)) {
ASSERT(heap_->new_space()->FromSpaceOffsetForAddress(new_addr) <=
heap_->new_space()->ToSpaceOffsetForAddress(old_addr));
if (heap()->new_space()->FromSpaceContains(new_addr)) {
ASSERT(heap()->new_space()->FromSpaceOffsetForAddress(new_addr) <=
heap()->new_space()->ToSpaceOffsetForAddress(old_addr));
} else {
ASSERT(heap_->TargetSpace(obj) == heap_->old_pointer_space() ||
heap_->TargetSpace(obj) == heap_->old_data_space());
ASSERT(heap()->TargetSpace(obj) == heap()->old_pointer_space() ||
heap()->TargetSpace(obj) == heap()->old_data_space());
}
#endif
// New and old addresses cannot overlap.
if (heap_->InNewSpace(HeapObject::FromAddress(new_addr))) {
heap_->CopyBlock(new_addr, old_addr, obj_size);
if (heap()->InNewSpace(HeapObject::FromAddress(new_addr))) {
heap()->CopyBlock(new_addr, old_addr, obj_size);
} else {
heap_->CopyBlockToOldSpaceAndUpdateRegionMarks(new_addr,
heap()->CopyBlockToOldSpaceAndUpdateRegionMarks(new_addr,
old_addr,
obj_size);
}
@ -3030,10 +3040,10 @@ int MarkCompactCollector::RelocateNewObject(HeapObject* obj) {
HeapObject* copied_to = HeapObject::FromAddress(new_addr);
if (copied_to->IsSharedFunctionInfo()) {
PROFILE(heap_->isolate(),
PROFILE(heap()->isolate(),
SharedFunctionInfoMoveEvent(old_addr, new_addr));
}
HEAP_PROFILE(heap_, ObjectMoveEvent(old_addr, new_addr));
HEAP_PROFILE(heap(), ObjectMoveEvent(old_addr, new_addr));
return obj_size;
}
@ -3042,7 +3052,7 @@ int MarkCompactCollector::RelocateNewObject(HeapObject* obj) {
void MarkCompactCollector::EnableCodeFlushing(bool enable) {
if (enable) {
if (code_flusher_ != NULL) return;
code_flusher_ = new CodeFlusher(heap_->isolate());
code_flusher_ = new CodeFlusher(heap()->isolate());
} else {
if (code_flusher_ == NULL) return;
delete code_flusher_;

View File

@ -98,8 +98,6 @@ class MarkingStack {
// -------------------------------------------------------------------------
// Mark-Compact collector
//
// All methods are static.
class OverflowedObjectsScanner;

View File

@ -393,9 +393,9 @@ void RelocInfo::Visit(Heap* heap) {
StaticVisitor::VisitPointer(heap, target_object_address());
CPU::FlushICache(pc_, sizeof(Address));
} else if (RelocInfo::IsCodeTarget(mode)) {
StaticVisitor::VisitCodeTarget(this);
StaticVisitor::VisitCodeTarget(heap, this);
} else if (mode == RelocInfo::GLOBAL_PROPERTY_CELL) {
StaticVisitor::VisitGlobalPropertyCell(this);
StaticVisitor::VisitGlobalPropertyCell(heap, this);
} else if (mode == RelocInfo::EXTERNAL_REFERENCE) {
StaticVisitor::VisitExternalReference(target_reference_address());
CPU::FlushICache(pc_, sizeof(Address));
@ -405,7 +405,7 @@ void RelocInfo::Visit(Heap* heap) {
IsPatchedReturnSequence()) ||
(RelocInfo::IsDebugBreakSlot(mode) &&
IsPatchedDebugBreakSlotSequence()))) {
StaticVisitor::VisitDebugTarget(this);
StaticVisitor::VisitDebugTarget(heap, this);
#endif
} else if (mode == RelocInfo::RUNTIME_ENTRY) {
StaticVisitor::VisitRuntimeEntry(this);