// Copyright 2016 the V8 project authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. #include "src/api.h" #include "src/heap/array-buffer-tracker.h" #include "src/heap/spaces.h" #include "src/isolate.h" #include "src/objects-inl.h" #include "test/cctest/cctest.h" #include "test/cctest/heap/heap-utils.h" namespace { typedef i::LocalArrayBufferTracker LocalTracker; bool IsTracked(i::JSArrayBuffer* buf) { return i::ArrayBufferTracker::IsTracked(buf); } } // namespace namespace v8 { namespace internal { namespace heap { // The following tests make sure that JSArrayBuffer tracking works expected when // moving the objects through various spaces during GC phases. TEST(ArrayBuffer_OnlyMC) { ManualGCScope manual_gc_scope; CcTest::InitializeVM(); LocalContext env; v8::Isolate* isolate = env->GetIsolate(); Heap* heap = reinterpret_cast(isolate)->heap(); JSArrayBuffer* raw_ab = nullptr; { v8::HandleScope handle_scope(isolate); Local ab = v8::ArrayBuffer::New(isolate, 100); Handle buf = v8::Utils::OpenHandle(*ab); CHECK(IsTracked(*buf)); heap::GcAndSweep(heap, OLD_SPACE); CHECK(IsTracked(*buf)); heap::GcAndSweep(heap, OLD_SPACE); CHECK(IsTracked(*buf)); raw_ab = *buf; // Prohibit page from being released. Page::FromAddress(buf->address())->MarkNeverEvacuate(); } // 2 GCs are needed because we promote to old space as live, meaning that // we will survive one GC. heap::GcAndSweep(heap, OLD_SPACE); heap::GcAndSweep(heap, OLD_SPACE); CHECK(!IsTracked(raw_ab)); } TEST(ArrayBuffer_OnlyScavenge) { ManualGCScope manual_gc_scope; CcTest::InitializeVM(); LocalContext env; v8::Isolate* isolate = env->GetIsolate(); Heap* heap = reinterpret_cast(isolate)->heap(); JSArrayBuffer* raw_ab = nullptr; { v8::HandleScope handle_scope(isolate); Local ab = v8::ArrayBuffer::New(isolate, 100); Handle buf = v8::Utils::OpenHandle(*ab); CHECK(IsTracked(*buf)); heap::GcAndSweep(heap, NEW_SPACE); CHECK(IsTracked(*buf)); heap::GcAndSweep(heap, NEW_SPACE); CHECK(IsTracked(*buf)); heap::GcAndSweep(heap, NEW_SPACE); CHECK(IsTracked(*buf)); raw_ab = *buf; // Prohibit page from being released. Page::FromAddress(buf->address())->MarkNeverEvacuate(); } // 2 GCs are needed because we promote to old space as live, meaning that // we will survive one GC. heap::GcAndSweep(heap, OLD_SPACE); heap::GcAndSweep(heap, OLD_SPACE); CHECK(!IsTracked(raw_ab)); } TEST(ArrayBuffer_ScavengeAndMC) { ManualGCScope manual_gc_scope; CcTest::InitializeVM(); LocalContext env; v8::Isolate* isolate = env->GetIsolate(); Heap* heap = reinterpret_cast(isolate)->heap(); JSArrayBuffer* raw_ab = nullptr; { v8::HandleScope handle_scope(isolate); Local ab = v8::ArrayBuffer::New(isolate, 100); Handle buf = v8::Utils::OpenHandle(*ab); CHECK(IsTracked(*buf)); heap::GcAndSweep(heap, NEW_SPACE); CHECK(IsTracked(*buf)); heap::GcAndSweep(heap, NEW_SPACE); CHECK(IsTracked(*buf)); heap::GcAndSweep(heap, OLD_SPACE); CHECK(IsTracked(*buf)); heap::GcAndSweep(heap, NEW_SPACE); CHECK(IsTracked(*buf)); raw_ab = *buf; // Prohibit page from being released. Page::FromAddress(buf->address())->MarkNeverEvacuate(); } // 2 GCs are needed because we promote to old space as live, meaning that // we will survive one GC. heap::GcAndSweep(heap, OLD_SPACE); heap::GcAndSweep(heap, OLD_SPACE); CHECK(!IsTracked(raw_ab)); } TEST(ArrayBuffer_Compaction) { if (FLAG_never_compact) return; ManualGCScope manual_gc_scope; FLAG_manual_evacuation_candidates_selection = true; CcTest::InitializeVM(); LocalContext env; v8::Isolate* isolate = env->GetIsolate(); Heap* heap = reinterpret_cast(isolate)->heap(); heap::AbandonCurrentlyFreeMemory(heap->old_space()); v8::HandleScope handle_scope(isolate); Local ab1 = v8::ArrayBuffer::New(isolate, 100); Handle buf1 = v8::Utils::OpenHandle(*ab1); CHECK(IsTracked(*buf1)); heap::GcAndSweep(heap, NEW_SPACE); heap::GcAndSweep(heap, NEW_SPACE); Page* page_before_gc = Page::FromAddress(buf1->address()); heap::ForceEvacuationCandidate(page_before_gc); CHECK(IsTracked(*buf1)); CcTest::CollectAllGarbage(); Page* page_after_gc = Page::FromAddress(buf1->address()); CHECK(IsTracked(*buf1)); CHECK_NE(page_before_gc, page_after_gc); } TEST(ArrayBuffer_UnregisterDuringSweep) { // Regular pages in old space (without compaction) are processed concurrently // in the sweeper. If we happen to unregister a buffer (either explicitly, or // implicitly through e.g. |Externalize|) we need to sync with the sweeper // task. // // Note: This test will will only fail on TSAN configurations. // Disable verify-heap since it forces sweeping to be completed in the // epilogue of the GC. #ifdef VERIFY_HEAP i::FLAG_verify_heap = false; #endif // VERIFY_HEAP ManualGCScope manual_gc_scope; CcTest::InitializeVM(); LocalContext env; v8::Isolate* isolate = env->GetIsolate(); Heap* heap = reinterpret_cast(isolate)->heap(); { v8::HandleScope handle_scope(isolate); Local ab = v8::ArrayBuffer::New(isolate, 100); Handle buf = v8::Utils::OpenHandle(*ab); { v8::HandleScope handle_scope(isolate); // Allocate another buffer on the same page to force processing a // non-empty set of buffers in the last GC. Local ab2 = v8::ArrayBuffer::New(isolate, 100); Handle buf2 = v8::Utils::OpenHandle(*ab2); CHECK(IsTracked(*buf)); CHECK(IsTracked(*buf)); heap::GcAndSweep(heap, NEW_SPACE); CHECK(IsTracked(*buf)); CHECK(IsTracked(*buf)); heap::GcAndSweep(heap, NEW_SPACE); CHECK(IsTracked(*buf)); CHECK(IsTracked(*buf2)); } CcTest::CollectGarbage(OLD_SPACE); // |Externalize| will cause the buffer to be |Unregister|ed. Without // barriers and proper synchronization this will trigger a data race on // TSAN. v8::ArrayBuffer::Contents contents = ab->Externalize(); heap->isolate()->array_buffer_allocator()->Free(contents.Data(), contents.ByteLength()); } } TEST(ArrayBuffer_NonLivePromotion) { if (!FLAG_incremental_marking) return; ManualGCScope manual_gc_scope; // The test verifies that the marking state is preserved when promoting // a buffer to old space. CcTest::InitializeVM(); LocalContext env; v8::Isolate* isolate = env->GetIsolate(); Heap* heap = reinterpret_cast(isolate)->heap(); JSArrayBuffer* raw_ab = nullptr; { v8::HandleScope handle_scope(isolate); Handle root = heap->isolate()->factory()->NewFixedArray(1, TENURED); { v8::HandleScope handle_scope(isolate); Local ab = v8::ArrayBuffer::New(isolate, 100); Handle buf = v8::Utils::OpenHandle(*ab); root->set(0, *buf); // Buffer that should not be promoted as live. } heap::SimulateIncrementalMarking(heap, false); CHECK(IsTracked(JSArrayBuffer::cast(root->get(0)))); heap::GcAndSweep(heap, NEW_SPACE); CHECK(IsTracked(JSArrayBuffer::cast(root->get(0)))); heap::GcAndSweep(heap, NEW_SPACE); CHECK(IsTracked(JSArrayBuffer::cast(root->get(0)))); raw_ab = JSArrayBuffer::cast(root->get(0)); root->set(0, heap->undefined_value()); heap::SimulateIncrementalMarking(heap, true); // Prohibit page from being released. Page::FromAddress(raw_ab->address())->MarkNeverEvacuate(); heap::GcAndSweep(heap, OLD_SPACE); CHECK(!IsTracked(raw_ab)); } } TEST(ArrayBuffer_LivePromotion) { if (!FLAG_incremental_marking) return; ManualGCScope manual_gc_scope; // The test verifies that the marking state is preserved when promoting // a buffer to old space. CcTest::InitializeVM(); LocalContext env; v8::Isolate* isolate = env->GetIsolate(); Heap* heap = reinterpret_cast(isolate)->heap(); JSArrayBuffer* raw_ab = nullptr; { v8::HandleScope handle_scope(isolate); Handle root = heap->isolate()->factory()->NewFixedArray(1, TENURED); { v8::HandleScope handle_scope(isolate); Local ab = v8::ArrayBuffer::New(isolate, 100); Handle buf = v8::Utils::OpenHandle(*ab); root->set(0, *buf); // Buffer that should be promoted as live. } heap::SimulateIncrementalMarking(heap, true); CHECK(IsTracked(JSArrayBuffer::cast(root->get(0)))); heap::GcAndSweep(heap, NEW_SPACE); CHECK(IsTracked(JSArrayBuffer::cast(root->get(0)))); heap::GcAndSweep(heap, NEW_SPACE); CHECK(IsTracked(JSArrayBuffer::cast(root->get(0)))); raw_ab = JSArrayBuffer::cast(root->get(0)); root->set(0, heap->undefined_value()); // Prohibit page from being released. Page::FromAddress(raw_ab->address())->MarkNeverEvacuate(); heap::GcAndSweep(heap, OLD_SPACE); CHECK(IsTracked(raw_ab)); } } TEST(ArrayBuffer_SemiSpaceCopyThenPagePromotion) { if (!i::FLAG_incremental_marking) return; ManualGCScope manual_gc_scope; // The test verifies that the marking state is preserved across semispace // copy. CcTest::InitializeVM(); LocalContext env; v8::Isolate* isolate = env->GetIsolate(); Heap* heap = reinterpret_cast(isolate)->heap(); heap::SealCurrentObjects(heap); { v8::HandleScope handle_scope(isolate); Handle root = heap->isolate()->factory()->NewFixedArray(1, TENURED); { v8::HandleScope handle_scope(isolate); Local ab = v8::ArrayBuffer::New(isolate, 100); Handle buf = v8::Utils::OpenHandle(*ab); root->set(0, *buf); // Buffer that should be promoted as live. Page::FromAddress(buf->address())->MarkNeverEvacuate(); } std::vector> handles; // Make the whole page transition from new->old, getting the buffers // processed in the sweeper (relying on marking information) instead of // processing during newspace evacuation. heap::FillCurrentPage(heap->new_space(), &handles); CHECK(IsTracked(JSArrayBuffer::cast(root->get(0)))); heap::GcAndSweep(heap, NEW_SPACE); heap::SimulateIncrementalMarking(heap, true); heap::GcAndSweep(heap, OLD_SPACE); CHECK(IsTracked(JSArrayBuffer::cast(root->get(0)))); } } UNINITIALIZED_TEST(ArrayBuffer_SemiSpaceCopyMultipleTasks) { if (FLAG_optimize_for_size) return; // Test allocates JSArrayBuffer on different pages before triggering a // full GC that performs the semispace copy. If parallelized, this test // ensures proper synchronization in TSAN configurations. FLAG_min_semi_space_size = 2 * Page::kPageSize / MB; v8::Isolate::CreateParams create_params; create_params.array_buffer_allocator = CcTest::array_buffer_allocator(); v8::Isolate* isolate = v8::Isolate::New(create_params); i::Isolate* i_isolate = reinterpret_cast(isolate); { v8::Isolate::Scope isolate_scope(isolate); v8::HandleScope handle_scope(isolate); v8::Context::New(isolate)->Enter(); Heap* heap = i_isolate->heap(); // Ensure heap is in a clean state. heap->CollectAllGarbage(Heap::kFinalizeIncrementalMarkingMask, GarbageCollectionReason::kTesting); heap->CollectAllGarbage(Heap::kFinalizeIncrementalMarkingMask, GarbageCollectionReason::kTesting); Local ab1 = v8::ArrayBuffer::New(isolate, 100); Handle buf1 = v8::Utils::OpenHandle(*ab1); heap::FillCurrentPage(heap->new_space()); Local ab2 = v8::ArrayBuffer::New(isolate, 100); Handle buf2 = v8::Utils::OpenHandle(*ab2); CHECK_NE(Page::FromAddress(buf1->address()), Page::FromAddress(buf2->address())); heap::GcAndSweep(heap, OLD_SPACE); } isolate->Dispose(); } TEST(ArrayBuffer_ExternalBackingStoreSizeIncreases) { CcTest::InitializeVM(); LocalContext env; v8::Isolate* isolate = env->GetIsolate(); Heap* heap = reinterpret_cast(isolate)->heap(); ExternalBackingStoreType type = ExternalBackingStoreType::kArrayBuffer; const size_t backing_store_before = heap->new_space()->ExternalBackingStoreBytes(type); { const size_t kArraybufferSize = 117; v8::HandleScope handle_scope(isolate); Local ab = v8::ArrayBuffer::New(isolate, kArraybufferSize); USE(ab); const size_t backing_store_after = heap->new_space()->ExternalBackingStoreBytes(type); CHECK_EQ(kArraybufferSize, backing_store_after - backing_store_before); } } TEST(ArrayBuffer_ExternalBackingStoreSizeDecreases) { CcTest::InitializeVM(); LocalContext env; v8::Isolate* isolate = env->GetIsolate(); Heap* heap = reinterpret_cast(isolate)->heap(); ExternalBackingStoreType type = ExternalBackingStoreType::kArrayBuffer; const size_t backing_store_before = heap->new_space()->ExternalBackingStoreBytes(type); { const size_t kArraybufferSize = 117; v8::HandleScope handle_scope(isolate); Local ab = v8::ArrayBuffer::New(isolate, kArraybufferSize); USE(ab); } heap::GcAndSweep(heap, OLD_SPACE); const size_t backing_store_after = heap->new_space()->ExternalBackingStoreBytes(type); CHECK_EQ(0, backing_store_after - backing_store_before); } TEST(ArrayBuffer_ExternalBackingStoreSizeIncreasesMarkCompact) { if (FLAG_never_compact) return; ManualGCScope manual_gc_scope; FLAG_manual_evacuation_candidates_selection = true; CcTest::InitializeVM(); LocalContext env; v8::Isolate* isolate = env->GetIsolate(); Heap* heap = reinterpret_cast(isolate)->heap(); heap::AbandonCurrentlyFreeMemory(heap->old_space()); ExternalBackingStoreType type = ExternalBackingStoreType::kArrayBuffer; const size_t backing_store_before = heap->old_space()->ExternalBackingStoreBytes(type); const size_t kArraybufferSize = 117; { v8::HandleScope handle_scope(isolate); Local ab1 = v8::ArrayBuffer::New(isolate, kArraybufferSize); Handle buf1 = v8::Utils::OpenHandle(*ab1); CHECK(IsTracked(*buf1)); heap::GcAndSweep(heap, NEW_SPACE); heap::GcAndSweep(heap, NEW_SPACE); Page* page_before_gc = Page::FromAddress(buf1->address()); heap::ForceEvacuationCandidate(page_before_gc); CHECK(IsTracked(*buf1)); CcTest::CollectAllGarbage(); const size_t backing_store_after = heap->old_space()->ExternalBackingStoreBytes(type); CHECK_EQ(kArraybufferSize, backing_store_after - backing_store_before); } heap::GcAndSweep(heap, OLD_SPACE); const size_t backing_store_after = heap->old_space()->ExternalBackingStoreBytes(type); CHECK_EQ(0, backing_store_after - backing_store_before); } } // namespace heap } // namespace internal } // namespace v8