c7cb9beca1
This is a reland offbfa9bf4ec
The arm64 was missing proper codegen for CFI, thus sizes were off. Original change's description: > Reland "[deoptimizer] Change deopt entries into builtins" > > This is a reland of7f58ced72e
> > It fixes the different exit size emitted on x64/Atom CPUs due to > performance tuning in TurboAssembler::Call. Additionally, add > cctests to verify the fixed size exits. > > Original change's description: > > [deoptimizer] Change deopt entries into builtins > > > > While the overall goal of this commit is to change deoptimization > > entries into builtins, there are multiple related things happening: > > > > - Deoptimization entries, formerly stubs (i.e. Code objects generated > > at runtime, guaranteed to be immovable), have been converted into > > builtins. The major restriction is that we now need to preserve the > > kRootRegister, which was formerly used on most architectures to pass > > the deoptimization id. The solution differs based on platform. > > - Renamed DEOPT_ENTRIES_OR_FOR_TESTING code kind to FOR_TESTING. > > - Removed heap/ support for immovable Code generation. > > - Removed the DeserializerData class (no longer needed). > > - arm64: to preserve 4-byte deopt exits, introduced a new optimization > > in which the final jump to the deoptimization entry is generated > > once per Code object, and deopt exits can continue to emit a > > near-call. > > - arm,ia32,x64: change to fixed-size deopt exits. This reduces exit > > sizes by 4/8, 5, and 5 bytes, respectively. > > > > On arm the deopt exit size is reduced from 12 (or 16) bytes to 8 bytes > > by using the same strategy as on arm64 (recalc deopt id from return > > address). Before: > > > > e300a002 movw r10, <id> > > e59fc024 ldr ip, [pc, <entry offset>] > > e12fff3c blx ip > > > > After: > > > > e59acb35 ldr ip, [r10, <entry offset>] > > e12fff3c blx ip > > > > On arm64 the deopt exit size remains 4 bytes (or 8 bytes in same cases > > with CFI). Additionally, up to 4 builtin jumps are emitted per Code > > object (max 32 bytes added overhead per Code object). Before: > > > > 9401cdae bl <entry offset> > > > > After: > > > > # eager deoptimization entry jump. > > f95b1f50 ldr x16, [x26, <eager entry offset>] > > d61f0200 br x16 > > # lazy deoptimization entry jump. > > f95b2b50 ldr x16, [x26, <lazy entry offset>] > > d61f0200 br x16 > > # the deopt exit. > > 97fffffc bl <eager deoptimization entry jump offset> > > > > On ia32 the deopt exit size is reduced from 10 to 5 bytes. Before: > > > > bb00000000 mov ebx,<id> > > e825f5372b call <entry> > > > > After: > > > > e8ea2256ba call <entry> > > > > On x64 the deopt exit size is reduced from 12 to 7 bytes. Before: > > > > 49c7c511000000 REX.W movq r13,<id> > > e8ea2f0700 call <entry> > > > > After: > > > > 41ff9560360000 call [r13+<entry offset>] > > > > Bug: v8:8661,v8:8768 > > Change-Id: I13e30aedc360474dc818fecc528ce87c3bfeed42 > > Reviewed-on: https://chromium-review.googlesource.com/c/v8/v8/+/2465834 > > Commit-Queue: Jakob Gruber <jgruber@chromium.org> > > Reviewed-by: Ross McIlroy <rmcilroy@chromium.org> > > Reviewed-by: Tobias Tebbi <tebbi@chromium.org> > > Reviewed-by: Ulan Degenbaev <ulan@chromium.org> > > Cr-Commit-Position: refs/heads/master@{#70597} > > Tbr: ulan@chromium.org, tebbi@chromium.org, rmcilroy@chromium.org > Bug: v8:8661,v8:8768,chromium:1140165 > Change-Id: Ibcd5c39c58a70bf2b2ac221aa375fc68d495e144 > Reviewed-on: https://chromium-review.googlesource.com/c/v8/v8/+/2485506 > Reviewed-by: Jakob Gruber <jgruber@chromium.org> > Reviewed-by: Tobias Tebbi <tebbi@chromium.org> > Commit-Queue: Jakob Gruber <jgruber@chromium.org> > Cr-Commit-Position: refs/heads/master@{#70655} Tbr: ulan@chromium.org, tebbi@chromium.org, rmcilroy@chromium.org Bug: v8:8661 Bug: v8:8768 Bug: chromium:1140165 Change-Id: I471cc94fc085e527dc9bfb5a84b96bd907c2333f Reviewed-on: https://chromium-review.googlesource.com/c/v8/v8/+/2488682 Reviewed-by: Jakob Gruber <jgruber@chromium.org> Commit-Queue: Jakob Gruber <jgruber@chromium.org> Cr-Commit-Position: refs/heads/master@{#70672}
341 lines
11 KiB
C++
341 lines
11 KiB
C++
// Copyright 2013 the V8 project authors. All rights reserved.
|
|
// Redistribution and use in source and binary forms, with or without
|
|
// modification, are permitted provided that the following conditions are
|
|
// met:
|
|
//
|
|
// * Redistributions of source code must retain the above copyright
|
|
// notice, this list of conditions and the following disclaimer.
|
|
// * Redistributions in binary form must reproduce the above
|
|
// copyright notice, this list of conditions and the following
|
|
// disclaimer in the documentation and/or other materials provided
|
|
// with the distribution.
|
|
// * Neither the name of Google Inc. nor the names of its
|
|
// contributors may be used to endorse or promote products derived
|
|
// from this software without specific prior written permission.
|
|
//
|
|
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
#include <stdlib.h>
|
|
|
|
#include "src/codegen/assembler-inl.h"
|
|
#include "src/codegen/macro-assembler.h"
|
|
#include "src/deoptimizer/deoptimizer.h"
|
|
#include "src/execution/simulator.h"
|
|
#include "src/init/v8.h"
|
|
#include "src/objects/objects-inl.h"
|
|
#include "src/utils/ostreams.h"
|
|
#include "test/cctest/cctest.h"
|
|
#include "test/common/assembler-tester.h"
|
|
|
|
namespace v8 {
|
|
namespace internal {
|
|
namespace test_macro_assembler_arm {
|
|
|
|
using F = void*(int x, int y, int p2, int p3, int p4);
|
|
|
|
#define __ masm->
|
|
|
|
using F3 = void*(void* p0, int p1, int p2, int p3, int p4);
|
|
using F5 = int(void*, void*, void*, void*, void*);
|
|
|
|
TEST(ExtractLane) {
|
|
if (!CpuFeatures::IsSupported(NEON)) return;
|
|
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope handles(isolate);
|
|
|
|
auto buffer = AllocateAssemblerBuffer();
|
|
MacroAssembler assembler(isolate, v8::internal::CodeObjectRequired::kYes,
|
|
buffer->CreateView());
|
|
MacroAssembler* masm = &assembler; // Create a pointer for the __ macro.
|
|
|
|
struct T {
|
|
int32_t i32x4_low[4];
|
|
int32_t i32x4_high[4];
|
|
int32_t i16x8_low[8];
|
|
int32_t i16x8_high[8];
|
|
int32_t i8x16_low[16];
|
|
int32_t i8x16_high[16];
|
|
int32_t f32x4_low[4];
|
|
int32_t f32x4_high[4];
|
|
int32_t i8x16_low_d[16];
|
|
int32_t i8x16_high_d[16];
|
|
};
|
|
T t;
|
|
|
|
__ stm(db_w, sp, r4.bit() | r5.bit() | lr.bit());
|
|
|
|
for (int i = 0; i < 4; i++) {
|
|
__ mov(r4, Operand(i));
|
|
__ vdup(Neon32, q1, r4);
|
|
__ ExtractLane(r5, q1, NeonS32, i);
|
|
__ str(r5, MemOperand(r0, offsetof(T, i32x4_low) + 4 * i));
|
|
SwVfpRegister si = SwVfpRegister::from_code(i);
|
|
__ ExtractLane(si, q1, i);
|
|
__ vstr(si, r0, offsetof(T, f32x4_low) + 4 * i);
|
|
}
|
|
|
|
for (int i = 0; i < 8; i++) {
|
|
__ mov(r4, Operand(i));
|
|
__ vdup(Neon16, q1, r4);
|
|
__ ExtractLane(r5, q1, NeonS16, i);
|
|
__ str(r5, MemOperand(r0, offsetof(T, i16x8_low) + 4 * i));
|
|
}
|
|
|
|
for (int i = 0; i < 16; i++) {
|
|
__ mov(r4, Operand(i));
|
|
__ vdup(Neon8, q1, r4);
|
|
__ ExtractLane(r5, q1, NeonS8, i);
|
|
__ str(r5, MemOperand(r0, offsetof(T, i8x16_low) + 4 * i));
|
|
}
|
|
|
|
for (int i = 0; i < 8; i++) {
|
|
__ mov(r4, Operand(i));
|
|
__ vdup(Neon8, q1, r4); // q1 = d2,d3
|
|
__ ExtractLane(r5, d2, NeonS8, i);
|
|
__ str(r5, MemOperand(r0, offsetof(T, i8x16_low_d) + 4 * i));
|
|
__ ExtractLane(r5, d3, NeonS8, i);
|
|
__ str(r5, MemOperand(r0, offsetof(T, i8x16_low_d) + 4 * (i + 8)));
|
|
}
|
|
|
|
if (CpuFeatures::IsSupported(VFP32DREGS)) {
|
|
for (int i = 0; i < 4; i++) {
|
|
__ mov(r4, Operand(-i));
|
|
__ vdup(Neon32, q15, r4);
|
|
__ ExtractLane(r5, q15, NeonS32, i);
|
|
__ str(r5, MemOperand(r0, offsetof(T, i32x4_high) + 4 * i));
|
|
SwVfpRegister si = SwVfpRegister::from_code(i);
|
|
__ ExtractLane(si, q15, i);
|
|
__ vstr(si, r0, offsetof(T, f32x4_high) + 4 * i);
|
|
}
|
|
|
|
for (int i = 0; i < 8; i++) {
|
|
__ mov(r4, Operand(-i));
|
|
__ vdup(Neon16, q15, r4);
|
|
__ ExtractLane(r5, q15, NeonS16, i);
|
|
__ str(r5, MemOperand(r0, offsetof(T, i16x8_high) + 4 * i));
|
|
}
|
|
|
|
for (int i = 0; i < 16; i++) {
|
|
__ mov(r4, Operand(-i));
|
|
__ vdup(Neon8, q15, r4);
|
|
__ ExtractLane(r5, q15, NeonS8, i);
|
|
__ str(r5, MemOperand(r0, offsetof(T, i8x16_high) + 4 * i));
|
|
}
|
|
|
|
for (int i = 0; i < 8; i++) {
|
|
__ mov(r4, Operand(-i));
|
|
__ vdup(Neon8, q15, r4); // q1 = d30,d31
|
|
__ ExtractLane(r5, d30, NeonS8, i);
|
|
__ str(r5, MemOperand(r0, offsetof(T, i8x16_high_d) + 4 * i));
|
|
__ ExtractLane(r5, d31, NeonS8, i);
|
|
__ str(r5, MemOperand(r0, offsetof(T, i8x16_high_d) + 4 * (i + 8)));
|
|
}
|
|
}
|
|
|
|
__ ldm(ia_w, sp, r4.bit() | r5.bit() | pc.bit());
|
|
|
|
CodeDesc desc;
|
|
masm->GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::FOR_TESTING).Build();
|
|
#ifdef DEBUG
|
|
StdoutStream os;
|
|
code->Print(os);
|
|
#endif
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
f.Call(&t, 0, 0, 0, 0);
|
|
for (int i = 0; i < 4; i++) {
|
|
CHECK_EQ(i, t.i32x4_low[i]);
|
|
CHECK_EQ(i, t.f32x4_low[i]);
|
|
}
|
|
for (int i = 0; i < 8; i++) {
|
|
CHECK_EQ(i, t.i16x8_low[i]);
|
|
}
|
|
for (int i = 0; i < 16; i++) {
|
|
CHECK_EQ(i, t.i8x16_low[i]);
|
|
}
|
|
for (int i = 0; i < 8; i++) {
|
|
CHECK_EQ(i, t.i8x16_low_d[i]);
|
|
CHECK_EQ(i, t.i8x16_low_d[i + 8]);
|
|
}
|
|
if (CpuFeatures::IsSupported(VFP32DREGS)) {
|
|
for (int i = 0; i < 4; i++) {
|
|
CHECK_EQ(-i, t.i32x4_high[i]);
|
|
CHECK_EQ(-i, t.f32x4_high[i]);
|
|
}
|
|
for (int i = 0; i < 8; i++) {
|
|
CHECK_EQ(-i, t.i16x8_high[i]);
|
|
}
|
|
for (int i = 0; i < 16; i++) {
|
|
CHECK_EQ(-i, t.i8x16_high[i]);
|
|
}
|
|
for (int i = 0; i < 8; i++) {
|
|
CHECK_EQ(-i, t.i8x16_high_d[i]);
|
|
CHECK_EQ(-i, t.i8x16_high_d[i + 8]);
|
|
}
|
|
}
|
|
}
|
|
|
|
TEST(ReplaceLane) {
|
|
if (!CpuFeatures::IsSupported(NEON)) return;
|
|
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope handles(isolate);
|
|
|
|
auto buffer = AllocateAssemblerBuffer();
|
|
MacroAssembler assembler(isolate, v8::internal::CodeObjectRequired::kYes,
|
|
buffer->CreateView());
|
|
MacroAssembler* masm = &assembler; // Create a pointer for the __ macro.
|
|
|
|
struct T {
|
|
int32_t i32x4_low[4];
|
|
int32_t i32x4_high[4];
|
|
int16_t i16x8_low[8];
|
|
int16_t i16x8_high[8];
|
|
int8_t i8x16_low[16];
|
|
int8_t i8x16_high[16];
|
|
int32_t f32x4_low[4];
|
|
int32_t f32x4_high[4];
|
|
};
|
|
T t;
|
|
|
|
__ stm(db_w, sp, r4.bit() | r5.bit() | r6.bit() | r7.bit() | lr.bit());
|
|
|
|
__ veor(q0, q0, q0); // Zero
|
|
__ veor(q1, q1, q1); // Zero
|
|
for (int i = 0; i < 4; i++) {
|
|
__ mov(r4, Operand(i));
|
|
__ ReplaceLane(q0, q0, r4, NeonS32, i);
|
|
SwVfpRegister si = SwVfpRegister::from_code(i);
|
|
__ vmov(si, r4);
|
|
__ ReplaceLane(q1, q1, si, i);
|
|
}
|
|
__ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, i32x4_low))));
|
|
__ vst1(Neon8, NeonListOperand(q0), NeonMemOperand(r4));
|
|
__ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, f32x4_low))));
|
|
__ vst1(Neon8, NeonListOperand(q1), NeonMemOperand(r4));
|
|
|
|
__ veor(q0, q0, q0); // Zero
|
|
for (int i = 0; i < 8; i++) {
|
|
__ mov(r4, Operand(i));
|
|
__ ReplaceLane(q0, q0, r4, NeonS16, i);
|
|
}
|
|
__ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, i16x8_low))));
|
|
__ vst1(Neon8, NeonListOperand(q0), NeonMemOperand(r4));
|
|
|
|
__ veor(q0, q0, q0); // Zero
|
|
for (int i = 0; i < 16; i++) {
|
|
__ mov(r4, Operand(i));
|
|
__ ReplaceLane(q0, q0, r4, NeonS8, i);
|
|
}
|
|
__ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, i8x16_low))));
|
|
__ vst1(Neon8, NeonListOperand(q0), NeonMemOperand(r4));
|
|
|
|
if (CpuFeatures::IsSupported(VFP32DREGS)) {
|
|
__ veor(q14, q14, q14); // Zero
|
|
__ veor(q15, q15, q15); // Zero
|
|
for (int i = 0; i < 4; i++) {
|
|
__ mov(r4, Operand(-i));
|
|
__ ReplaceLane(q14, q14, r4, NeonS32, i);
|
|
SwVfpRegister si = SwVfpRegister::from_code(i);
|
|
__ vmov(si, r4);
|
|
__ ReplaceLane(q15, q15, si, i);
|
|
}
|
|
__ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, i32x4_high))));
|
|
__ vst1(Neon8, NeonListOperand(q14), NeonMemOperand(r4));
|
|
__ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, f32x4_high))));
|
|
__ vst1(Neon8, NeonListOperand(q15), NeonMemOperand(r4));
|
|
|
|
__ veor(q14, q14, q14); // Zero
|
|
for (int i = 0; i < 8; i++) {
|
|
__ mov(r4, Operand(-i));
|
|
__ ReplaceLane(q14, q14, r4, NeonS16, i);
|
|
}
|
|
__ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, i16x8_high))));
|
|
__ vst1(Neon8, NeonListOperand(q14), NeonMemOperand(r4));
|
|
|
|
__ veor(q14, q14, q14); // Zero
|
|
for (int i = 0; i < 16; i++) {
|
|
__ mov(r4, Operand(-i));
|
|
__ ReplaceLane(q14, q14, r4, NeonS8, i);
|
|
}
|
|
__ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, i8x16_high))));
|
|
__ vst1(Neon8, NeonListOperand(q14), NeonMemOperand(r4));
|
|
}
|
|
|
|
__ ldm(ia_w, sp, r4.bit() | r5.bit() | r6.bit() | r7.bit() | pc.bit());
|
|
|
|
CodeDesc desc;
|
|
masm->GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::FOR_TESTING).Build();
|
|
#ifdef DEBUG
|
|
StdoutStream os;
|
|
code->Print(os);
|
|
#endif
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
f.Call(&t, 0, 0, 0, 0);
|
|
for (int i = 0; i < 4; i++) {
|
|
CHECK_EQ(i, t.i32x4_low[i]);
|
|
CHECK_EQ(i, t.f32x4_low[i]);
|
|
}
|
|
for (int i = 0; i < 8; i++) {
|
|
CHECK_EQ(i, t.i16x8_low[i]);
|
|
}
|
|
for (int i = 0; i < 16; i++) {
|
|
CHECK_EQ(i, t.i8x16_low[i]);
|
|
}
|
|
if (CpuFeatures::IsSupported(VFP32DREGS)) {
|
|
for (int i = 0; i < 4; i++) {
|
|
CHECK_EQ(-i, t.i32x4_high[i]);
|
|
CHECK_EQ(-i, t.f32x4_high[i]);
|
|
}
|
|
for (int i = 0; i < 8; i++) {
|
|
CHECK_EQ(-i, t.i16x8_high[i]);
|
|
}
|
|
for (int i = 0; i < 16; i++) {
|
|
CHECK_EQ(-i, t.i8x16_high[i]);
|
|
}
|
|
}
|
|
}
|
|
|
|
TEST(DeoptExitSizeIsFixed) {
|
|
CHECK(Deoptimizer::kSupportsFixedDeoptExitSizes);
|
|
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope handles(isolate);
|
|
auto buffer = AllocateAssemblerBuffer();
|
|
MacroAssembler masm(isolate, v8::internal::CodeObjectRequired::kYes,
|
|
buffer->CreateView());
|
|
|
|
STATIC_ASSERT(static_cast<int>(kFirstDeoptimizeKind) == 0);
|
|
for (int i = 0; i < kDeoptimizeKindCount; i++) {
|
|
DeoptimizeKind kind = static_cast<DeoptimizeKind>(i);
|
|
Builtins::Name target = Deoptimizer::GetDeoptimizationEntry(isolate, kind);
|
|
Label before_exit;
|
|
masm.bind(&before_exit);
|
|
masm.CallForDeoptimization(target, 42, &before_exit, kind, &before_exit);
|
|
CHECK_EQ(masm.SizeOfCodeGeneratedSince(&before_exit),
|
|
kind == DeoptimizeKind::kLazy
|
|
? Deoptimizer::kLazyDeoptExitSize
|
|
: Deoptimizer::kNonLazyDeoptExitSize);
|
|
}
|
|
}
|
|
|
|
#undef __
|
|
|
|
} // namespace test_macro_assembler_arm
|
|
} // namespace internal
|
|
} // namespace v8
|