2015-06-15 08:32:52 +00:00
|
|
|
// Copyright 2014 the V8 project authors. All rights reserved.
|
|
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
|
|
// found in the LICENSE file.
|
|
|
|
|
|
|
|
#include <cmath>
|
2016-12-19 12:50:30 +00:00
|
|
|
#include <iostream>
|
2015-06-15 08:32:52 +00:00
|
|
|
#include <limits>
|
|
|
|
|
2019-05-22 12:44:24 +00:00
|
|
|
#include "src/handles/handles-inl.h"
|
2015-06-15 08:32:52 +00:00
|
|
|
#include "src/heap/heap.h"
|
2019-02-14 21:10:30 +00:00
|
|
|
#include "src/heap/spaces-inl.h"
|
2019-05-23 08:51:46 +00:00
|
|
|
#include "src/objects/objects-inl.h"
|
2017-07-13 17:13:53 +00:00
|
|
|
#include "test/unittests/test-utils.h"
|
2015-06-15 08:32:52 +00:00
|
|
|
#include "testing/gtest/include/gtest/gtest.h"
|
|
|
|
|
|
|
|
namespace v8 {
|
|
|
|
namespace internal {
|
|
|
|
|
2019-03-29 09:47:48 +00:00
|
|
|
using HeapTest = TestWithIsolate;
|
|
|
|
using HeapWithPointerCompressionTest = TestWithIsolateAndPointerCompression;
|
2017-07-13 17:13:53 +00:00
|
|
|
|
2019-06-05 21:08:15 +00:00
|
|
|
TEST(Heap, YoungGenerationSizeFromOldGenerationSize) {
|
2017-07-19 09:53:06 +00:00
|
|
|
const size_t MB = static_cast<size_t>(i::MB);
|
2019-05-28 15:51:07 +00:00
|
|
|
const size_t KB = static_cast<size_t>(i::KB);
|
2017-07-19 09:53:06 +00:00
|
|
|
const size_t pm = i::Heap::kPointerMultiplier;
|
2019-06-05 21:08:15 +00:00
|
|
|
ASSERT_EQ(3 * 512u * pm * KB,
|
|
|
|
i::Heap::YoungGenerationSizeFromOldGenerationSize(128u * pm * MB));
|
|
|
|
ASSERT_EQ(3 * 2048u * pm * KB,
|
|
|
|
i::Heap::YoungGenerationSizeFromOldGenerationSize(256u * pm * MB));
|
|
|
|
ASSERT_EQ(3 * 4096u * pm * KB,
|
|
|
|
i::Heap::YoungGenerationSizeFromOldGenerationSize(512u * pm * MB));
|
|
|
|
ASSERT_EQ(3 * 8192u * pm * KB,
|
|
|
|
i::Heap::YoungGenerationSizeFromOldGenerationSize(1024u * pm * MB));
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST(Heap, GenerationSizesFromHeapSize) {
|
|
|
|
const size_t MB = static_cast<size_t>(i::MB);
|
|
|
|
const size_t KB = static_cast<size_t>(i::KB);
|
|
|
|
const size_t pm = i::Heap::kPointerMultiplier;
|
|
|
|
size_t old, young;
|
2019-05-28 15:51:07 +00:00
|
|
|
|
2019-06-05 21:08:15 +00:00
|
|
|
i::Heap::GenerationSizesFromHeapSize(1 * KB, &young, &old);
|
|
|
|
ASSERT_EQ(0u, old);
|
|
|
|
ASSERT_EQ(0u, young);
|
2019-05-28 15:51:07 +00:00
|
|
|
|
2019-06-05 21:08:15 +00:00
|
|
|
i::Heap::GenerationSizesFromHeapSize(1 * KB + 3 * 512u * pm * KB, &young,
|
|
|
|
&old);
|
|
|
|
ASSERT_EQ(1 * KB, old);
|
|
|
|
ASSERT_EQ(3 * 512u * pm * KB, young);
|
2019-05-28 15:51:07 +00:00
|
|
|
|
2019-06-05 21:08:15 +00:00
|
|
|
i::Heap::GenerationSizesFromHeapSize(128 * pm * MB + 3 * 512 * pm * KB,
|
|
|
|
&young, &old);
|
|
|
|
ASSERT_EQ(128u * pm * MB, old);
|
|
|
|
ASSERT_EQ(3 * 512u * pm * KB, young);
|
2019-05-28 15:51:07 +00:00
|
|
|
|
2019-06-05 21:08:15 +00:00
|
|
|
i::Heap::GenerationSizesFromHeapSize(256u * pm * MB + 3 * 2048 * pm * KB,
|
|
|
|
&young, &old);
|
|
|
|
ASSERT_EQ(256u * pm * MB, old);
|
|
|
|
ASSERT_EQ(3 * 2048u * pm * KB, young);
|
2019-05-28 15:51:07 +00:00
|
|
|
|
2019-06-05 21:08:15 +00:00
|
|
|
i::Heap::GenerationSizesFromHeapSize(512u * pm * MB + 3 * 4096 * pm * KB,
|
|
|
|
&young, &old);
|
|
|
|
ASSERT_EQ(512u * pm * MB, old);
|
|
|
|
ASSERT_EQ(3 * 4096u * pm * KB, young);
|
|
|
|
|
|
|
|
i::Heap::GenerationSizesFromHeapSize(1024u * pm * MB + 3 * 8192 * pm * KB,
|
|
|
|
&young, &old);
|
|
|
|
ASSERT_EQ(1024u * pm * MB, old);
|
|
|
|
ASSERT_EQ(3 * 8192u * pm * KB, young);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST(Heap, HeapSizeFromPhysicalMemory) {
|
|
|
|
const size_t MB = static_cast<size_t>(i::MB);
|
|
|
|
const size_t pm = i::Heap::kPointerMultiplier;
|
2019-05-28 15:51:07 +00:00
|
|
|
|
2019-06-05 21:08:15 +00:00
|
|
|
// The expected value is old_generation_size + 3 * semi_space_size.
|
|
|
|
ASSERT_EQ(128 * pm * MB + 3 * 512 * pm * KB,
|
|
|
|
i::Heap::HeapSizeFromPhysicalMemory(0u));
|
|
|
|
ASSERT_EQ(128 * pm * MB + 3 * 512 * pm * KB,
|
|
|
|
i::Heap::HeapSizeFromPhysicalMemory(512u * MB));
|
|
|
|
ASSERT_EQ(256 * pm * MB + 3 * 2048 * pm * KB,
|
|
|
|
i::Heap::HeapSizeFromPhysicalMemory(1024u * MB));
|
|
|
|
ASSERT_EQ(512 * pm * MB + 3 * 4096 * pm * KB,
|
|
|
|
i::Heap::HeapSizeFromPhysicalMemory(2048u * MB));
|
|
|
|
ASSERT_EQ(
|
|
|
|
1024 * pm * MB + 3 * 8192 * pm * KB,
|
|
|
|
i::Heap::HeapSizeFromPhysicalMemory(static_cast<uint64_t>(4096u) * MB));
|
|
|
|
ASSERT_EQ(
|
|
|
|
1024 * pm * MB + 3 * 8192 * pm * KB,
|
|
|
|
i::Heap::HeapSizeFromPhysicalMemory(static_cast<uint64_t>(8192u) * MB));
|
2017-05-23 17:00:57 +00:00
|
|
|
}
|
|
|
|
|
2017-07-13 17:13:53 +00:00
|
|
|
TEST_F(HeapTest, ASLR) {
|
|
|
|
#if V8_TARGET_ARCH_X64
|
|
|
|
#if V8_OS_MACOSX
|
|
|
|
Heap* heap = i_isolate()->heap();
|
|
|
|
std::set<void*> hints;
|
|
|
|
for (int i = 0; i < 1000; i++) {
|
|
|
|
hints.insert(heap->GetRandomMmapAddr());
|
|
|
|
}
|
|
|
|
if (hints.size() == 1) {
|
|
|
|
EXPECT_TRUE((*hints.begin()) == nullptr);
|
2017-12-15 17:59:57 +00:00
|
|
|
EXPECT_TRUE(i::GetRandomMmapAddr() == nullptr);
|
2017-07-13 17:13:53 +00:00
|
|
|
} else {
|
|
|
|
// It is unlikely that 1000 random samples will collide to less then 500
|
|
|
|
// values.
|
|
|
|
EXPECT_GT(hints.size(), 500u);
|
|
|
|
const uintptr_t kRegionMask = 0xFFFFFFFFu;
|
|
|
|
void* first = *hints.begin();
|
|
|
|
for (void* hint : hints) {
|
|
|
|
uintptr_t diff = reinterpret_cast<uintptr_t>(first) ^
|
|
|
|
reinterpret_cast<uintptr_t>(hint);
|
|
|
|
EXPECT_LE(diff, kRegionMask);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
#endif // V8_OS_MACOSX
|
|
|
|
#endif // V8_TARGET_ARCH_X64
|
|
|
|
}
|
|
|
|
|
2018-09-06 14:04:07 +00:00
|
|
|
TEST_F(HeapTest, ExternalLimitDefault) {
|
|
|
|
Heap* heap = i_isolate()->heap();
|
2018-10-26 13:30:12 +00:00
|
|
|
EXPECT_EQ(kExternalAllocationSoftLimit,
|
|
|
|
heap->isolate()->isolate_data()->external_memory_limit_);
|
2018-09-06 14:04:07 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(HeapTest, ExternalLimitStaysAboveDefaultForExplicitHandling) {
|
|
|
|
v8_isolate()->AdjustAmountOfExternalAllocatedMemory(+10 * MB);
|
|
|
|
v8_isolate()->AdjustAmountOfExternalAllocatedMemory(-10 * MB);
|
|
|
|
Heap* heap = i_isolate()->heap();
|
2018-10-26 13:30:12 +00:00
|
|
|
EXPECT_GE(heap->isolate()->isolate_data()->external_memory_limit_,
|
|
|
|
kExternalAllocationSoftLimit);
|
2018-09-06 14:04:07 +00:00
|
|
|
}
|
|
|
|
|
2018-10-30 12:48:12 +00:00
|
|
|
#if V8_TARGET_ARCH_64_BIT
|
|
|
|
TEST_F(HeapWithPointerCompressionTest, HeapLayout) {
|
|
|
|
// Produce some garbage.
|
|
|
|
RunJS(
|
|
|
|
"let ar = [];"
|
|
|
|
"for (let i = 0; i < 100; i++) {"
|
|
|
|
" ar.push(Array(i));"
|
|
|
|
"}"
|
|
|
|
"ar.push(Array(32 * 1024 * 1024));");
|
|
|
|
|
|
|
|
Address isolate_root = i_isolate()->isolate_root();
|
|
|
|
EXPECT_TRUE(IsAligned(isolate_root, size_t{4} * GB));
|
|
|
|
|
|
|
|
// Check that all memory chunks belong this region.
|
2019-10-07 16:23:50 +00:00
|
|
|
base::AddressRegion heap_reservation(isolate_root, size_t{4} * GB);
|
2018-10-30 12:48:12 +00:00
|
|
|
|
2018-11-21 19:19:02 +00:00
|
|
|
OldGenerationMemoryChunkIterator iter(i_isolate()->heap());
|
2018-10-30 12:48:12 +00:00
|
|
|
for (;;) {
|
|
|
|
MemoryChunk* chunk = iter.next();
|
|
|
|
if (chunk == nullptr) break;
|
|
|
|
|
|
|
|
Address address = chunk->address();
|
|
|
|
size_t size = chunk->area_end() - address;
|
|
|
|
EXPECT_TRUE(heap_reservation.contains(address, size));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
#endif // V8_TARGET_ARCH_64_BIT
|
|
|
|
|
2015-06-15 08:32:52 +00:00
|
|
|
} // namespace internal
|
|
|
|
} // namespace v8
|