c51041f454
With the new Turbofan variants (NCI and Turboprop), we need a way to distinguish between them both during and after compilation. We initially introduced CompilationTarget to track the variant during compilation, but decided to reuse the code kind as the canonical spot to store this information instead. Why? Because it is an established mechanism, already available in most of the necessary spots (inside the pipeline, on Code objects, in profiling traces). This CL removes CompilationTarget and adds a new NATIVE_CONTEXT_INDEPENDENT kind, plus helper functions to determine various things about a given code kind (e.g.: does this code kind deopt?). As a (very large) drive-by, refactor both Code::Kind and AbstractCode::Kind into a new CodeKind enum class. Bug: v8:8888 Change-Id: Ie858b9a53311b0731630be35cf5cd108dee95b39 Reviewed-on: https://chromium-review.googlesource.com/c/v8/v8/+/2336793 Commit-Queue: Jakob Gruber <jgruber@chromium.org> Reviewed-by: Clemens Backes <clemensb@chromium.org> Reviewed-by: Ross McIlroy <rmcilroy@chromium.org> Reviewed-by: Dominik Inführ <dinfuehr@chromium.org> Reviewed-by: Georg Neis <neis@chromium.org> Cr-Commit-Position: refs/heads/master@{#69244}
10443 lines
366 KiB
C++
10443 lines
366 KiB
C++
// Copyright 2012 the V8 project authors. All rights reserved.
|
|
// Redistribution and use in source and binary forms, with or without
|
|
// modification, are permitted provided that the following conditions are
|
|
// met:
|
|
//
|
|
// * Redistributions of source code must retain the above copyright
|
|
// notice, this list of conditions and the following disclaimer.
|
|
// * Redistributions in binary form must reproduce the above
|
|
// copyright notice, this list of conditions and the following
|
|
// disclaimer in the documentation and/or other materials provided
|
|
// with the distribution.
|
|
// * Neither the name of Google Inc. nor the names of its
|
|
// contributors may be used to endorse or promote products derived
|
|
// from this software without specific prior written permission.
|
|
//
|
|
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
#include <iostream> // NOLINT(readability/streams)
|
|
|
|
#include "src/init/v8.h"
|
|
|
|
#include "src/base/utils/random-number-generator.h"
|
|
#include "src/codegen/assembler-inl.h"
|
|
#include "src/codegen/macro-assembler.h"
|
|
#include "src/diagnostics/disassembler.h"
|
|
#include "src/execution/simulator.h"
|
|
#include "src/heap/factory.h"
|
|
|
|
#include "test/cctest/cctest.h"
|
|
|
|
namespace v8 {
|
|
namespace internal {
|
|
|
|
// Define these function prototypes to match JSEntryFunction in execution.cc.
|
|
// TODO(mips): Refine these signatures per test case.
|
|
using F1 = void*(int x, int p1, int p2, int p3, int p4);
|
|
using F2 = void*(int x, int y, int p2, int p3, int p4);
|
|
using F3 = void*(void* p, int p1, int p2, int p3, int p4);
|
|
using F4 = void*(void* p0, void* p1, int p2, int p3, int p4);
|
|
|
|
#define __ assm.
|
|
|
|
TEST(MIPS0) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
// Addition.
|
|
__ addu(v0, a0, a1);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F2>::FromCode(*code);
|
|
int res = reinterpret_cast<int>(f.Call(0xAB0, 0xC, 0, 0, 0));
|
|
CHECK_EQ(static_cast<int32_t>(0xABC), res);
|
|
}
|
|
|
|
|
|
TEST(MIPS1) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
Label L, C;
|
|
|
|
__ mov(a1, a0);
|
|
__ li(v0, 0);
|
|
__ b(&C);
|
|
__ nop();
|
|
|
|
__ bind(&L);
|
|
__ addu(v0, v0, a1);
|
|
__ addiu(a1, a1, -1);
|
|
|
|
__ bind(&C);
|
|
__ xori(v1, a1, 0);
|
|
__ Branch(&L, ne, v1, Operand(0));
|
|
__ nop();
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F1>::FromCode(*code);
|
|
int res = reinterpret_cast<int>(f.Call(50, 0, 0, 0, 0));
|
|
CHECK_EQ(1275, res);
|
|
}
|
|
|
|
|
|
TEST(MIPS2) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
Label exit, error;
|
|
|
|
// ----- Test all instructions.
|
|
|
|
// Test lui, ori, and addiu, used in the li pseudo-instruction.
|
|
// This way we can then safely load registers with chosen values.
|
|
|
|
__ ori(t0, zero_reg, 0);
|
|
__ lui(t0, 0x1234);
|
|
__ ori(t0, t0, 0);
|
|
__ ori(t0, t0, 0x0F0F);
|
|
__ ori(t0, t0, 0xF0F0);
|
|
__ addiu(t1, t0, 1);
|
|
__ addiu(t2, t1, -0x10);
|
|
|
|
// Load values in temporary registers.
|
|
__ li(t0, 0x00000004);
|
|
__ li(t1, 0x00001234);
|
|
__ li(t2, 0x12345678);
|
|
__ li(t3, 0x7FFFFFFF);
|
|
__ li(t4, 0xFFFFFFFC);
|
|
__ li(t5, 0xFFFFEDCC);
|
|
__ li(t6, 0xEDCBA988);
|
|
__ li(t7, 0x80000000);
|
|
|
|
// SPECIAL class.
|
|
__ srl(v0, t2, 8); // 0x00123456
|
|
__ sll(v0, v0, 11); // 0x91A2B000
|
|
__ sra(v0, v0, 3); // 0xF2345600
|
|
__ srav(v0, v0, t0); // 0xFF234560
|
|
__ sllv(v0, v0, t0); // 0xF2345600
|
|
__ srlv(v0, v0, t0); // 0x0F234560
|
|
__ Branch(&error, ne, v0, Operand(0x0F234560));
|
|
__ nop();
|
|
|
|
__ addu(v0, t0, t1); // 0x00001238
|
|
__ subu(v0, v0, t0); // 0x00001234
|
|
__ Branch(&error, ne, v0, Operand(0x00001234));
|
|
__ nop();
|
|
__ addu(v1, t3, t0);
|
|
__ Branch(&error, ne, v1, Operand(0x80000003));
|
|
__ nop();
|
|
__ subu(v1, t7, t0); // 0x7FFFFFFC
|
|
__ Branch(&error, ne, v1, Operand(0x7FFFFFFC));
|
|
__ nop();
|
|
|
|
__ and_(v0, t1, t2); // 0x00001230
|
|
__ or_(v0, v0, t1); // 0x00001234
|
|
__ xor_(v0, v0, t2); // 0x1234444C
|
|
__ nor(v0, v0, t2); // 0xEDCBA987
|
|
__ Branch(&error, ne, v0, Operand(0xEDCBA983));
|
|
__ nop();
|
|
|
|
__ slt(v0, t7, t3);
|
|
__ Branch(&error, ne, v0, Operand(0x1));
|
|
__ nop();
|
|
__ sltu(v0, t7, t3);
|
|
__ Branch(&error, ne, v0, Operand(zero_reg));
|
|
__ nop();
|
|
// End of SPECIAL class.
|
|
|
|
__ addiu(v0, zero_reg, 0x7421); // 0x00007421
|
|
__ addiu(v0, v0, -0x1); // 0x00007420
|
|
__ addiu(v0, v0, -0x20); // 0x00007400
|
|
__ Branch(&error, ne, v0, Operand(0x00007400));
|
|
__ nop();
|
|
__ addiu(v1, t3, 0x1); // 0x80000000
|
|
__ Branch(&error, ne, v1, Operand(0x80000000));
|
|
__ nop();
|
|
|
|
__ slti(v0, t1, 0x00002000); // 0x1
|
|
__ slti(v0, v0, 0xFFFF8000); // 0x0
|
|
__ Branch(&error, ne, v0, Operand(zero_reg));
|
|
__ nop();
|
|
__ sltiu(v0, t1, 0x00002000); // 0x1
|
|
__ sltiu(v0, v0, 0x00008000); // 0x1
|
|
__ Branch(&error, ne, v0, Operand(0x1));
|
|
__ nop();
|
|
|
|
__ andi(v0, t1, 0xF0F0); // 0x00001030
|
|
__ ori(v0, v0, 0x8A00); // 0x00009A30
|
|
__ xori(v0, v0, 0x83CC); // 0x000019FC
|
|
__ Branch(&error, ne, v0, Operand(0x000019FC));
|
|
__ nop();
|
|
__ lui(v1, 0x8123); // 0x81230000
|
|
__ Branch(&error, ne, v1, Operand(0x81230000));
|
|
__ nop();
|
|
|
|
// Bit twiddling instructions & conditional moves.
|
|
// Uses t0-t7 as set above.
|
|
__ Clz(v0, t0); // 29
|
|
__ Clz(v1, t1); // 19
|
|
__ addu(v0, v0, v1); // 48
|
|
__ Clz(v1, t2); // 3
|
|
__ addu(v0, v0, v1); // 51
|
|
__ Clz(v1, t7); // 0
|
|
__ addu(v0, v0, v1); // 51
|
|
__ Branch(&error, ne, v0, Operand(51));
|
|
__ Movn(a0, t3, t0); // Move a0<-t3 (t0 is NOT 0).
|
|
__ Ins(a0, t1, 12, 8); // 0x7FF34FFF
|
|
__ Branch(&error, ne, a0, Operand(0x7FF34FFF));
|
|
__ Movz(a0, t6, t7); // a0 not updated (t7 is NOT 0).
|
|
__ Ext(a1, a0, 8, 12); // 0x34F
|
|
__ Branch(&error, ne, a1, Operand(0x34F));
|
|
__ Movz(a0, t6, v1); // a0<-t6, v0 is 0, from 8 instr back.
|
|
__ Branch(&error, ne, a0, Operand(t6));
|
|
|
|
// Everything was correctly executed. Load the expected result.
|
|
__ li(v0, 0x31415926);
|
|
__ b(&exit);
|
|
__ nop();
|
|
|
|
__ bind(&error);
|
|
// Got an error. Return a wrong result.
|
|
__ li(v0, 666);
|
|
|
|
__ bind(&exit);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F2>::FromCode(*code);
|
|
int res = reinterpret_cast<int>(f.Call(0xAB0, 0xC, 0, 0, 0));
|
|
CHECK_EQ(static_cast<int32_t>(0x31415926), res);
|
|
}
|
|
|
|
|
|
TEST(MIPS3) {
|
|
// Test floating point instructions.
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
struct T {
|
|
double a;
|
|
double b;
|
|
double c;
|
|
double d;
|
|
double e;
|
|
double f;
|
|
double g;
|
|
double h;
|
|
double i;
|
|
float fa;
|
|
float fb;
|
|
float fc;
|
|
float fd;
|
|
float fe;
|
|
float ff;
|
|
float fg;
|
|
};
|
|
T t;
|
|
|
|
// Create a function that accepts &t, and loads, manipulates, and stores
|
|
// the doubles t.a ... t.f.
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
// Double precision floating point instructions.
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(T, a)));
|
|
__ Ldc1(f6, MemOperand(a0, offsetof(T, b)));
|
|
__ add_d(f8, f4, f6);
|
|
__ Sdc1(f8, MemOperand(a0, offsetof(T, c))); // c = a + b.
|
|
|
|
__ mov_d(f10, f8); // c
|
|
__ neg_d(f12, f6); // -b
|
|
__ sub_d(f10, f10, f12);
|
|
__ Sdc1(f10, MemOperand(a0, offsetof(T, d))); // d = c - (-b).
|
|
|
|
__ Sdc1(f4, MemOperand(a0, offsetof(T, b))); // b = a.
|
|
|
|
__ li(t0, 120);
|
|
__ mtc1(t0, f14);
|
|
__ cvt_d_w(f14, f14); // f14 = 120.0.
|
|
__ mul_d(f10, f10, f14);
|
|
__ Sdc1(f10, MemOperand(a0, offsetof(T, e))); // e = d * 120 = 1.8066e16.
|
|
|
|
__ div_d(f12, f10, f4);
|
|
__ Sdc1(f12, MemOperand(a0, offsetof(T, f))); // f = e / a = 120.44.
|
|
|
|
__ sqrt_d(f14, f12);
|
|
__ Sdc1(f14, MemOperand(a0, offsetof(T, g)));
|
|
// g = sqrt(f) = 10.97451593465515908537
|
|
|
|
if (IsMipsArchVariant(kMips32r2)) {
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(T, h)));
|
|
__ Ldc1(f6, MemOperand(a0, offsetof(T, i)));
|
|
__ madd_d(f14, f6, f4, f6);
|
|
__ Sdc1(f14, MemOperand(a0, offsetof(T, h)));
|
|
}
|
|
|
|
// Single precision floating point instructions.
|
|
__ lwc1(f4, MemOperand(a0, offsetof(T, fa)) );
|
|
__ lwc1(f6, MemOperand(a0, offsetof(T, fb)) );
|
|
__ add_s(f8, f4, f6);
|
|
__ swc1(f8, MemOperand(a0, offsetof(T, fc)) ); // fc = fa + fb.
|
|
|
|
__ neg_s(f10, f6); // -fb
|
|
__ sub_s(f10, f8, f10);
|
|
__ swc1(f10, MemOperand(a0, offsetof(T, fd)) ); // fd = fc - (-fb).
|
|
|
|
__ swc1(f4, MemOperand(a0, offsetof(T, fb)) ); // fb = fa.
|
|
|
|
__ li(t0, 120);
|
|
__ mtc1(t0, f14);
|
|
__ cvt_s_w(f14, f14); // f14 = 120.0.
|
|
__ mul_s(f10, f10, f14);
|
|
__ swc1(f10, MemOperand(a0, offsetof(T, fe)) ); // fe = fd * 120
|
|
|
|
__ div_s(f12, f10, f4);
|
|
__ swc1(f12, MemOperand(a0, offsetof(T, ff)) ); // ff = fe / fa
|
|
|
|
__ sqrt_s(f14, f12);
|
|
__ swc1(f14, MemOperand(a0, offsetof(T, fg)) );
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
// Double test values.
|
|
t.a = 1.5e14;
|
|
t.b = 2.75e11;
|
|
t.c = 0.0;
|
|
t.d = 0.0;
|
|
t.e = 0.0;
|
|
t.f = 0.0;
|
|
t.h = 1.5;
|
|
t.i = 2.75;
|
|
// Single test values.
|
|
t.fa = 1.5e6;
|
|
t.fb = 2.75e4;
|
|
t.fc = 0.0;
|
|
t.fd = 0.0;
|
|
t.fe = 0.0;
|
|
t.ff = 0.0;
|
|
f.Call(&t, 0, 0, 0, 0);
|
|
// Expected double results.
|
|
CHECK_EQ(1.5e14, t.a);
|
|
CHECK_EQ(1.5e14, t.b);
|
|
CHECK_EQ(1.50275e14, t.c);
|
|
CHECK_EQ(1.50550e14, t.d);
|
|
CHECK_EQ(1.8066e16, t.e);
|
|
CHECK_EQ(120.44, t.f);
|
|
CHECK_EQ(10.97451593465515908537, t.g);
|
|
if (IsMipsArchVariant(kMips32r2)) {
|
|
CHECK_EQ(6.875, t.h);
|
|
}
|
|
// Expected single results.
|
|
CHECK_EQ(1.5e6, t.fa);
|
|
CHECK_EQ(1.5e6, t.fb);
|
|
CHECK_EQ(1.5275e06, t.fc);
|
|
CHECK_EQ(1.5550e06, t.fd);
|
|
CHECK_EQ(1.866e08, t.fe);
|
|
CHECK_EQ(124.40000152587890625, t.ff);
|
|
CHECK_EQ(11.1534748077392578125, t.fg);
|
|
}
|
|
|
|
|
|
TEST(MIPS4) {
|
|
// Exchange between GP anf FP registers is done through memory
|
|
// on FPXX compiled binaries and architectures that do not support
|
|
// MTHC1 and MTFC1. If this is the case, skipping this test.
|
|
if (IsFpxxMode() &&
|
|
(IsMipsArchVariant(kMips32r1) || IsMipsArchVariant(kLoongson))) {
|
|
return;
|
|
}
|
|
|
|
// Test moves between floating point and integer registers.
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
struct T {
|
|
double a;
|
|
double b;
|
|
double c;
|
|
};
|
|
T t;
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(T, a)));
|
|
__ Ldc1(f6, MemOperand(a0, offsetof(T, b)));
|
|
|
|
// Swap f4 and f6, by using four integer registers, t0-t3.
|
|
if (IsFp32Mode()) {
|
|
__ mfc1(t0, f4);
|
|
__ mfc1(t1, f5);
|
|
__ mfc1(t2, f6);
|
|
__ mfc1(t3, f7);
|
|
|
|
__ mtc1(t0, f6);
|
|
__ mtc1(t1, f7);
|
|
__ mtc1(t2, f4);
|
|
__ mtc1(t3, f5);
|
|
} else {
|
|
CHECK(!IsMipsArchVariant(kMips32r1) && !IsMipsArchVariant(kLoongson));
|
|
DCHECK(IsFp64Mode() || IsFpxxMode());
|
|
__ mfc1(t0, f4);
|
|
__ mfhc1(t1, f4);
|
|
__ mfc1(t2, f6);
|
|
__ mfhc1(t3, f6);
|
|
|
|
__ mtc1(t0, f6);
|
|
__ mthc1(t1, f6);
|
|
__ mtc1(t2, f4);
|
|
__ mthc1(t3, f4);
|
|
}
|
|
|
|
// Store the swapped f4 and f5 back to memory.
|
|
__ Sdc1(f4, MemOperand(a0, offsetof(T, a)));
|
|
__ Sdc1(f6, MemOperand(a0, offsetof(T, c)));
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
t.a = 1.5e22;
|
|
t.b = 2.75e11;
|
|
t.c = 17.17;
|
|
f.Call(&t, 0, 0, 0, 0);
|
|
|
|
CHECK_EQ(2.75e11, t.a);
|
|
CHECK_EQ(2.75e11, t.b);
|
|
CHECK_EQ(1.5e22, t.c);
|
|
}
|
|
|
|
|
|
TEST(MIPS5) {
|
|
// Test conversions between doubles and integers.
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
struct T {
|
|
double a;
|
|
double b;
|
|
int i;
|
|
int j;
|
|
};
|
|
T t;
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
// Load all structure elements to registers.
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(T, a)));
|
|
__ Ldc1(f6, MemOperand(a0, offsetof(T, b)));
|
|
__ lw(t0, MemOperand(a0, offsetof(T, i)) );
|
|
__ lw(t1, MemOperand(a0, offsetof(T, j)) );
|
|
|
|
// Convert double in f4 to int in element i.
|
|
__ cvt_w_d(f8, f4);
|
|
__ mfc1(t2, f8);
|
|
__ sw(t2, MemOperand(a0, offsetof(T, i)) );
|
|
|
|
// Convert double in f6 to int in element j.
|
|
__ cvt_w_d(f10, f6);
|
|
__ mfc1(t3, f10);
|
|
__ sw(t3, MemOperand(a0, offsetof(T, j)) );
|
|
|
|
// Convert int in original i (t0) to double in a.
|
|
__ mtc1(t0, f12);
|
|
__ cvt_d_w(f0, f12);
|
|
__ Sdc1(f0, MemOperand(a0, offsetof(T, a)));
|
|
|
|
// Convert int in original j (t1) to double in b.
|
|
__ mtc1(t1, f14);
|
|
__ cvt_d_w(f2, f14);
|
|
__ Sdc1(f2, MemOperand(a0, offsetof(T, b)));
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
t.a = 1.5e4;
|
|
t.b = 2.75e8;
|
|
t.i = 12345678;
|
|
t.j = -100000;
|
|
f.Call(&t, 0, 0, 0, 0);
|
|
|
|
CHECK_EQ(12345678.0, t.a);
|
|
CHECK_EQ(-100000.0, t.b);
|
|
CHECK_EQ(15000, t.i);
|
|
CHECK_EQ(275000000, t.j);
|
|
}
|
|
|
|
|
|
TEST(MIPS6) {
|
|
// Test simple memory loads and stores.
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
struct T {
|
|
uint32_t ui;
|
|
int32_t si;
|
|
int32_t r1;
|
|
int32_t r2;
|
|
int32_t r3;
|
|
int32_t r4;
|
|
int32_t r5;
|
|
int32_t r6;
|
|
};
|
|
T t;
|
|
|
|
Assembler assm(AssemblerOptions{});
|
|
|
|
// Basic word load/store.
|
|
__ lw(t0, MemOperand(a0, offsetof(T, ui)) );
|
|
__ sw(t0, MemOperand(a0, offsetof(T, r1)) );
|
|
|
|
// lh with positive data.
|
|
__ lh(t1, MemOperand(a0, offsetof(T, ui)) );
|
|
__ sw(t1, MemOperand(a0, offsetof(T, r2)) );
|
|
|
|
// lh with negative data.
|
|
__ lh(t2, MemOperand(a0, offsetof(T, si)) );
|
|
__ sw(t2, MemOperand(a0, offsetof(T, r3)) );
|
|
|
|
// lhu with negative data.
|
|
__ lhu(t3, MemOperand(a0, offsetof(T, si)) );
|
|
__ sw(t3, MemOperand(a0, offsetof(T, r4)) );
|
|
|
|
// lb with negative data.
|
|
__ lb(t4, MemOperand(a0, offsetof(T, si)) );
|
|
__ sw(t4, MemOperand(a0, offsetof(T, r5)) );
|
|
|
|
// sh writes only 1/2 of word.
|
|
__ lui(t5, 0x3333);
|
|
__ ori(t5, t5, 0x3333);
|
|
__ sw(t5, MemOperand(a0, offsetof(T, r6)) );
|
|
__ lhu(t5, MemOperand(a0, offsetof(T, si)) );
|
|
__ sh(t5, MemOperand(a0, offsetof(T, r6)) );
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
t.ui = 0x11223344;
|
|
t.si = 0x99AABBCC;
|
|
f.Call(&t, 0, 0, 0, 0);
|
|
|
|
CHECK_EQ(static_cast<int32_t>(0x11223344), t.r1);
|
|
#if __BYTE_ORDER == __LITTLE_ENDIAN
|
|
CHECK_EQ(static_cast<int32_t>(0x3344), t.r2);
|
|
CHECK_EQ(static_cast<int32_t>(0xFFFFBBCC), t.r3);
|
|
CHECK_EQ(static_cast<int32_t>(0x0000BBCC), t.r4);
|
|
CHECK_EQ(static_cast<int32_t>(0xFFFFFFCC), t.r5);
|
|
CHECK_EQ(static_cast<int32_t>(0x3333BBCC), t.r6);
|
|
#elif __BYTE_ORDER == __BIG_ENDIAN
|
|
CHECK_EQ(static_cast<int32_t>(0x1122), t.r2);
|
|
CHECK_EQ(static_cast<int32_t>(0xFFFF99AA), t.r3);
|
|
CHECK_EQ(static_cast<int32_t>(0x000099AA), t.r4);
|
|
CHECK_EQ(static_cast<int32_t>(0xFFFFFF99), t.r5);
|
|
CHECK_EQ(static_cast<int32_t>(0x99AA3333), t.r6);
|
|
#else
|
|
#error Unknown endianness
|
|
#endif
|
|
}
|
|
|
|
|
|
TEST(MIPS7) {
|
|
// Test floating point compare and branch instructions.
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
struct T {
|
|
double a;
|
|
double b;
|
|
double c;
|
|
double d;
|
|
double e;
|
|
double f;
|
|
int32_t result;
|
|
};
|
|
T t;
|
|
|
|
// Create a function that accepts &t, and loads, manipulates, and stores
|
|
// the doubles t.a ... t.f.
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
Label neither_is_nan, less_than, outa_here;
|
|
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(T, a)));
|
|
__ Ldc1(f6, MemOperand(a0, offsetof(T, b)));
|
|
if (!IsMipsArchVariant(kMips32r6)) {
|
|
__ c(UN, D, f4, f6);
|
|
__ bc1f(&neither_is_nan);
|
|
} else {
|
|
__ cmp(UN, L, f2, f4, f6);
|
|
__ bc1eqz(&neither_is_nan, f2);
|
|
}
|
|
__ nop();
|
|
__ sw(zero_reg, MemOperand(a0, offsetof(T, result)) );
|
|
__ Branch(&outa_here);
|
|
|
|
__ bind(&neither_is_nan);
|
|
|
|
if (IsMipsArchVariant(kLoongson)) {
|
|
__ c(OLT, D, f6, f4);
|
|
__ bc1t(&less_than);
|
|
} else if (IsMipsArchVariant(kMips32r6)) {
|
|
__ cmp(OLT, L, f2, f6, f4);
|
|
__ bc1nez(&less_than, f2);
|
|
} else {
|
|
__ c(OLT, D, f6, f4, 2);
|
|
__ bc1t(&less_than, 2);
|
|
}
|
|
|
|
__ nop();
|
|
__ sw(zero_reg, MemOperand(a0, offsetof(T, result)) );
|
|
__ Branch(&outa_here);
|
|
|
|
__ bind(&less_than);
|
|
__ Addu(t0, zero_reg, Operand(1));
|
|
__ sw(t0, MemOperand(a0, offsetof(T, result)) ); // Set true.
|
|
|
|
|
|
// This test-case should have additional tests.
|
|
|
|
__ bind(&outa_here);
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
t.a = 1.5e14;
|
|
t.b = 2.75e11;
|
|
t.c = 2.0;
|
|
t.d = -4.0;
|
|
t.e = 0.0;
|
|
t.f = 0.0;
|
|
t.result = 0;
|
|
f.Call(&t, 0, 0, 0, 0);
|
|
CHECK_EQ(1.5e14, t.a);
|
|
CHECK_EQ(2.75e11, t.b);
|
|
CHECK_EQ(1, t.result);
|
|
}
|
|
|
|
|
|
TEST(MIPS8) {
|
|
// Test ROTR and ROTRV instructions.
|
|
if (IsMipsArchVariant(kMips32r2)) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
struct T {
|
|
int32_t input;
|
|
int32_t result_rotr_4;
|
|
int32_t result_rotr_8;
|
|
int32_t result_rotr_12;
|
|
int32_t result_rotr_16;
|
|
int32_t result_rotr_20;
|
|
int32_t result_rotr_24;
|
|
int32_t result_rotr_28;
|
|
int32_t result_rotrv_4;
|
|
int32_t result_rotrv_8;
|
|
int32_t result_rotrv_12;
|
|
int32_t result_rotrv_16;
|
|
int32_t result_rotrv_20;
|
|
int32_t result_rotrv_24;
|
|
int32_t result_rotrv_28;
|
|
};
|
|
T t;
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
// Basic word load.
|
|
__ lw(t0, MemOperand(a0, offsetof(T, input)) );
|
|
|
|
// ROTR instruction (called through the Ror macro).
|
|
__ Ror(t1, t0, 0x0004);
|
|
__ Ror(t2, t0, 0x0008);
|
|
__ Ror(t3, t0, 0x000C);
|
|
__ Ror(t4, t0, 0x0010);
|
|
__ Ror(t5, t0, 0x0014);
|
|
__ Ror(t6, t0, 0x0018);
|
|
__ Ror(t7, t0, 0x001C);
|
|
|
|
// Basic word store.
|
|
__ sw(t1, MemOperand(a0, offsetof(T, result_rotr_4)) );
|
|
__ sw(t2, MemOperand(a0, offsetof(T, result_rotr_8)) );
|
|
__ sw(t3, MemOperand(a0, offsetof(T, result_rotr_12)) );
|
|
__ sw(t4, MemOperand(a0, offsetof(T, result_rotr_16)) );
|
|
__ sw(t5, MemOperand(a0, offsetof(T, result_rotr_20)) );
|
|
__ sw(t6, MemOperand(a0, offsetof(T, result_rotr_24)) );
|
|
__ sw(t7, MemOperand(a0, offsetof(T, result_rotr_28)) );
|
|
|
|
// ROTRV instruction (called through the Ror macro).
|
|
__ li(t7, 0x0004);
|
|
__ Ror(t1, t0, t7);
|
|
__ li(t7, 0x0008);
|
|
__ Ror(t2, t0, t7);
|
|
__ li(t7, 0x000C);
|
|
__ Ror(t3, t0, t7);
|
|
__ li(t7, 0x0010);
|
|
__ Ror(t4, t0, t7);
|
|
__ li(t7, 0x0014);
|
|
__ Ror(t5, t0, t7);
|
|
__ li(t7, 0x0018);
|
|
__ Ror(t6, t0, t7);
|
|
__ li(t7, 0x001C);
|
|
__ Ror(t7, t0, t7);
|
|
|
|
// Basic word store.
|
|
__ sw(t1, MemOperand(a0, offsetof(T, result_rotrv_4)) );
|
|
__ sw(t2, MemOperand(a0, offsetof(T, result_rotrv_8)) );
|
|
__ sw(t3, MemOperand(a0, offsetof(T, result_rotrv_12)) );
|
|
__ sw(t4, MemOperand(a0, offsetof(T, result_rotrv_16)) );
|
|
__ sw(t5, MemOperand(a0, offsetof(T, result_rotrv_20)) );
|
|
__ sw(t6, MemOperand(a0, offsetof(T, result_rotrv_24)) );
|
|
__ sw(t7, MemOperand(a0, offsetof(T, result_rotrv_28)) );
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
t.input = 0x12345678;
|
|
f.Call(&t, 0x0, 0, 0, 0);
|
|
CHECK_EQ(static_cast<int32_t>(0x81234567), t.result_rotr_4);
|
|
CHECK_EQ(static_cast<int32_t>(0x78123456), t.result_rotr_8);
|
|
CHECK_EQ(static_cast<int32_t>(0x67812345), t.result_rotr_12);
|
|
CHECK_EQ(static_cast<int32_t>(0x56781234), t.result_rotr_16);
|
|
CHECK_EQ(static_cast<int32_t>(0x45678123), t.result_rotr_20);
|
|
CHECK_EQ(static_cast<int32_t>(0x34567812), t.result_rotr_24);
|
|
CHECK_EQ(static_cast<int32_t>(0x23456781), t.result_rotr_28);
|
|
|
|
CHECK_EQ(static_cast<int32_t>(0x81234567), t.result_rotrv_4);
|
|
CHECK_EQ(static_cast<int32_t>(0x78123456), t.result_rotrv_8);
|
|
CHECK_EQ(static_cast<int32_t>(0x67812345), t.result_rotrv_12);
|
|
CHECK_EQ(static_cast<int32_t>(0x56781234), t.result_rotrv_16);
|
|
CHECK_EQ(static_cast<int32_t>(0x45678123), t.result_rotrv_20);
|
|
CHECK_EQ(static_cast<int32_t>(0x34567812), t.result_rotrv_24);
|
|
CHECK_EQ(static_cast<int32_t>(0x23456781), t.result_rotrv_28);
|
|
}
|
|
}
|
|
|
|
|
|
TEST(MIPS9) {
|
|
// Test BRANCH improvements.
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
Label exit, exit2, exit3;
|
|
|
|
__ Branch(&exit, ge, a0, Operand(zero_reg));
|
|
__ Branch(&exit2, ge, a0, Operand(0x00001FFF));
|
|
__ Branch(&exit3, ge, a0, Operand(0x0001FFFF));
|
|
|
|
__ bind(&exit);
|
|
__ bind(&exit2);
|
|
__ bind(&exit3);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
USE(code);
|
|
}
|
|
|
|
|
|
TEST(MIPS10) {
|
|
// Test conversions between doubles and words.
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
struct T {
|
|
double a;
|
|
double b;
|
|
int32_t dbl_mant;
|
|
int32_t dbl_exp;
|
|
int32_t word;
|
|
int32_t b_word;
|
|
};
|
|
T t;
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
if (IsMipsArchVariant(kMips32r1) || IsMipsArchVariant(kLoongson)) return;
|
|
|
|
// Load all structure elements to registers.
|
|
// (f0, f1) = a (fp32), f0 = a (fp64)
|
|
__ Ldc1(f0, MemOperand(a0, offsetof(T, a)));
|
|
|
|
__ mfc1(t0, f0); // t0 = f0(31..0)
|
|
__ mfhc1(t1, f0); // t1 = sign_extend(f0(63..32))
|
|
__ sw(t0, MemOperand(a0, offsetof(T, dbl_mant))); // dbl_mant = t0
|
|
__ sw(t1, MemOperand(a0, offsetof(T, dbl_exp))); // dbl_exp = t1
|
|
|
|
// Convert double in f0 to word, save hi/lo parts.
|
|
__ cvt_w_d(f0, f0); // a_word = (word)a
|
|
__ mfc1(t0, f0); // f0 has a 32-bits word. t0 = a_word
|
|
__ sw(t0, MemOperand(a0, offsetof(T, word))); // word = a_word
|
|
|
|
// Convert the b word to double b.
|
|
__ lw(t0, MemOperand(a0, offsetof(T, b_word)));
|
|
__ mtc1(t0, f8); // f8 has a 32-bits word.
|
|
__ cvt_d_w(f10, f8);
|
|
__ Sdc1(f10, MemOperand(a0, offsetof(T, b)));
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
t.a = 2.147483646e+09; // 0x7FFFFFFE -> 0xFF80000041DFFFFF as double.
|
|
t.b_word = 0x0FF00FF0; // 0x0FF00FF0 -> 0x as double.
|
|
f.Call(&t, 0, 0, 0, 0);
|
|
CHECK_EQ(static_cast<int32_t>(0x41DFFFFF), t.dbl_exp);
|
|
CHECK_EQ(static_cast<int32_t>(0xFF800000), t.dbl_mant);
|
|
CHECK_EQ(static_cast<int32_t>(0x7FFFFFFE), t.word);
|
|
// 0x0FF00FF0 -> 2.6739096+e08
|
|
CHECK_EQ(2.6739096e08, t.b);
|
|
}
|
|
|
|
|
|
TEST(MIPS11) {
|
|
// Do not run test on MIPS32r6, as these instructions are removed.
|
|
if (IsMipsArchVariant(kMips32r6)) return;
|
|
// Test LWL, LWR, SWL and SWR instructions.
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
struct T {
|
|
int32_t reg_init;
|
|
int32_t mem_init;
|
|
int32_t lwl_0;
|
|
int32_t lwl_1;
|
|
int32_t lwl_2;
|
|
int32_t lwl_3;
|
|
int32_t lwr_0;
|
|
int32_t lwr_1;
|
|
int32_t lwr_2;
|
|
int32_t lwr_3;
|
|
int32_t swl_0;
|
|
int32_t swl_1;
|
|
int32_t swl_2;
|
|
int32_t swl_3;
|
|
int32_t swr_0;
|
|
int32_t swr_1;
|
|
int32_t swr_2;
|
|
int32_t swr_3;
|
|
};
|
|
T t;
|
|
|
|
Assembler assm(AssemblerOptions{});
|
|
|
|
// Test all combinations of LWL and vAddr.
|
|
__ lw(t0, MemOperand(a0, offsetof(T, reg_init)) );
|
|
__ lwl(t0, MemOperand(a0, offsetof(T, mem_init)) );
|
|
__ sw(t0, MemOperand(a0, offsetof(T, lwl_0)) );
|
|
|
|
__ lw(t1, MemOperand(a0, offsetof(T, reg_init)) );
|
|
__ lwl(t1, MemOperand(a0, offsetof(T, mem_init) + 1) );
|
|
__ sw(t1, MemOperand(a0, offsetof(T, lwl_1)) );
|
|
|
|
__ lw(t2, MemOperand(a0, offsetof(T, reg_init)) );
|
|
__ lwl(t2, MemOperand(a0, offsetof(T, mem_init) + 2) );
|
|
__ sw(t2, MemOperand(a0, offsetof(T, lwl_2)) );
|
|
|
|
__ lw(t3, MemOperand(a0, offsetof(T, reg_init)) );
|
|
__ lwl(t3, MemOperand(a0, offsetof(T, mem_init) + 3) );
|
|
__ sw(t3, MemOperand(a0, offsetof(T, lwl_3)) );
|
|
|
|
// Test all combinations of LWR and vAddr.
|
|
__ lw(t0, MemOperand(a0, offsetof(T, reg_init)) );
|
|
__ lwr(t0, MemOperand(a0, offsetof(T, mem_init)) );
|
|
__ sw(t0, MemOperand(a0, offsetof(T, lwr_0)) );
|
|
|
|
__ lw(t1, MemOperand(a0, offsetof(T, reg_init)) );
|
|
__ lwr(t1, MemOperand(a0, offsetof(T, mem_init) + 1) );
|
|
__ sw(t1, MemOperand(a0, offsetof(T, lwr_1)) );
|
|
|
|
__ lw(t2, MemOperand(a0, offsetof(T, reg_init)) );
|
|
__ lwr(t2, MemOperand(a0, offsetof(T, mem_init) + 2) );
|
|
__ sw(t2, MemOperand(a0, offsetof(T, lwr_2)) );
|
|
|
|
__ lw(t3, MemOperand(a0, offsetof(T, reg_init)) );
|
|
__ lwr(t3, MemOperand(a0, offsetof(T, mem_init) + 3) );
|
|
__ sw(t3, MemOperand(a0, offsetof(T, lwr_3)) );
|
|
|
|
// Test all combinations of SWL and vAddr.
|
|
__ lw(t0, MemOperand(a0, offsetof(T, mem_init)) );
|
|
__ sw(t0, MemOperand(a0, offsetof(T, swl_0)) );
|
|
__ lw(t0, MemOperand(a0, offsetof(T, reg_init)) );
|
|
__ swl(t0, MemOperand(a0, offsetof(T, swl_0)) );
|
|
|
|
__ lw(t1, MemOperand(a0, offsetof(T, mem_init)) );
|
|
__ sw(t1, MemOperand(a0, offsetof(T, swl_1)) );
|
|
__ lw(t1, MemOperand(a0, offsetof(T, reg_init)) );
|
|
__ swl(t1, MemOperand(a0, offsetof(T, swl_1) + 1) );
|
|
|
|
__ lw(t2, MemOperand(a0, offsetof(T, mem_init)) );
|
|
__ sw(t2, MemOperand(a0, offsetof(T, swl_2)) );
|
|
__ lw(t2, MemOperand(a0, offsetof(T, reg_init)) );
|
|
__ swl(t2, MemOperand(a0, offsetof(T, swl_2) + 2) );
|
|
|
|
__ lw(t3, MemOperand(a0, offsetof(T, mem_init)) );
|
|
__ sw(t3, MemOperand(a0, offsetof(T, swl_3)) );
|
|
__ lw(t3, MemOperand(a0, offsetof(T, reg_init)) );
|
|
__ swl(t3, MemOperand(a0, offsetof(T, swl_3) + 3) );
|
|
|
|
// Test all combinations of SWR and vAddr.
|
|
__ lw(t0, MemOperand(a0, offsetof(T, mem_init)) );
|
|
__ sw(t0, MemOperand(a0, offsetof(T, swr_0)) );
|
|
__ lw(t0, MemOperand(a0, offsetof(T, reg_init)) );
|
|
__ swr(t0, MemOperand(a0, offsetof(T, swr_0)) );
|
|
|
|
__ lw(t1, MemOperand(a0, offsetof(T, mem_init)) );
|
|
__ sw(t1, MemOperand(a0, offsetof(T, swr_1)) );
|
|
__ lw(t1, MemOperand(a0, offsetof(T, reg_init)) );
|
|
__ swr(t1, MemOperand(a0, offsetof(T, swr_1) + 1) );
|
|
|
|
__ lw(t2, MemOperand(a0, offsetof(T, mem_init)) );
|
|
__ sw(t2, MemOperand(a0, offsetof(T, swr_2)) );
|
|
__ lw(t2, MemOperand(a0, offsetof(T, reg_init)) );
|
|
__ swr(t2, MemOperand(a0, offsetof(T, swr_2) + 2) );
|
|
|
|
__ lw(t3, MemOperand(a0, offsetof(T, mem_init)) );
|
|
__ sw(t3, MemOperand(a0, offsetof(T, swr_3)) );
|
|
__ lw(t3, MemOperand(a0, offsetof(T, reg_init)) );
|
|
__ swr(t3, MemOperand(a0, offsetof(T, swr_3) + 3) );
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
t.reg_init = 0xAABBCCDD;
|
|
t.mem_init = 0x11223344;
|
|
|
|
f.Call(&t, 0, 0, 0, 0);
|
|
|
|
#if __BYTE_ORDER == __LITTLE_ENDIAN
|
|
CHECK_EQ(static_cast<int32_t>(0x44BBCCDD), t.lwl_0);
|
|
CHECK_EQ(static_cast<int32_t>(0x3344CCDD), t.lwl_1);
|
|
CHECK_EQ(static_cast<int32_t>(0x223344DD), t.lwl_2);
|
|
CHECK_EQ(static_cast<int32_t>(0x11223344), t.lwl_3);
|
|
|
|
CHECK_EQ(static_cast<int32_t>(0x11223344), t.lwr_0);
|
|
CHECK_EQ(static_cast<int32_t>(0xAA112233), t.lwr_1);
|
|
CHECK_EQ(static_cast<int32_t>(0xAABB1122), t.lwr_2);
|
|
CHECK_EQ(static_cast<int32_t>(0xAABBCC11), t.lwr_3);
|
|
|
|
CHECK_EQ(static_cast<int32_t>(0x112233AA), t.swl_0);
|
|
CHECK_EQ(static_cast<int32_t>(0x1122AABB), t.swl_1);
|
|
CHECK_EQ(static_cast<int32_t>(0x11AABBCC), t.swl_2);
|
|
CHECK_EQ(static_cast<int32_t>(0xAABBCCDD), t.swl_3);
|
|
|
|
CHECK_EQ(static_cast<int32_t>(0xAABBCCDD), t.swr_0);
|
|
CHECK_EQ(static_cast<int32_t>(0xBBCCDD44), t.swr_1);
|
|
CHECK_EQ(static_cast<int32_t>(0xCCDD3344), t.swr_2);
|
|
CHECK_EQ(static_cast<int32_t>(0xDD223344), t.swr_3);
|
|
#elif __BYTE_ORDER == __BIG_ENDIAN
|
|
CHECK_EQ(static_cast<int32_t>(0x11223344), t.lwl_0);
|
|
CHECK_EQ(static_cast<int32_t>(0x223344DD), t.lwl_1);
|
|
CHECK_EQ(static_cast<int32_t>(0x3344CCDD), t.lwl_2);
|
|
CHECK_EQ(static_cast<int32_t>(0x44BBCCDD), t.lwl_3);
|
|
|
|
CHECK_EQ(static_cast<int32_t>(0xAABBCC11), t.lwr_0);
|
|
CHECK_EQ(static_cast<int32_t>(0xAABB1122), t.lwr_1);
|
|
CHECK_EQ(static_cast<int32_t>(0xAA112233), t.lwr_2);
|
|
CHECK_EQ(static_cast<int32_t>(0x11223344), t.lwr_3);
|
|
|
|
CHECK_EQ(static_cast<int32_t>(0xAABBCCDD), t.swl_0);
|
|
CHECK_EQ(static_cast<int32_t>(0x11AABBCC), t.swl_1);
|
|
CHECK_EQ(static_cast<int32_t>(0x1122AABB), t.swl_2);
|
|
CHECK_EQ(static_cast<int32_t>(0x112233AA), t.swl_3);
|
|
|
|
CHECK_EQ(static_cast<int32_t>(0xDD223344), t.swr_0);
|
|
CHECK_EQ(static_cast<int32_t>(0xCCDD3344), t.swr_1);
|
|
CHECK_EQ(static_cast<int32_t>(0xBBCCDD44), t.swr_2);
|
|
CHECK_EQ(static_cast<int32_t>(0xAABBCCDD), t.swr_3);
|
|
#else
|
|
#error Unknown endianness
|
|
#endif
|
|
}
|
|
|
|
|
|
TEST(MIPS12) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
struct T {
|
|
int32_t x;
|
|
int32_t y;
|
|
int32_t y1;
|
|
int32_t y2;
|
|
int32_t y3;
|
|
int32_t y4;
|
|
};
|
|
T t;
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
__ mov(t6, fp); // Save frame pointer.
|
|
__ mov(fp, a0); // Access struct T by fp.
|
|
__ lw(t0, MemOperand(a0, offsetof(T, y)) );
|
|
__ lw(t3, MemOperand(a0, offsetof(T, y4)) );
|
|
|
|
__ addu(t1, t0, t3);
|
|
__ subu(t4, t0, t3);
|
|
__ nop();
|
|
__ push(t0); // These instructions disappear after opt.
|
|
__ Pop();
|
|
__ addu(t0, t0, t0);
|
|
__ nop();
|
|
__ Pop(); // These instructions disappear after opt.
|
|
__ push(t3);
|
|
__ nop();
|
|
__ push(t3); // These instructions disappear after opt.
|
|
__ pop(t3);
|
|
__ nop();
|
|
__ push(t3);
|
|
__ pop(t4);
|
|
__ nop();
|
|
__ sw(t0, MemOperand(fp, offsetof(T, y)) );
|
|
__ lw(t0, MemOperand(fp, offsetof(T, y)) );
|
|
__ nop();
|
|
__ sw(t0, MemOperand(fp, offsetof(T, y)) );
|
|
__ lw(t1, MemOperand(fp, offsetof(T, y)) );
|
|
__ nop();
|
|
__ push(t1);
|
|
__ lw(t1, MemOperand(fp, offsetof(T, y)) );
|
|
__ pop(t1);
|
|
__ nop();
|
|
__ push(t1);
|
|
__ lw(t2, MemOperand(fp, offsetof(T, y)) );
|
|
__ pop(t1);
|
|
__ nop();
|
|
__ push(t1);
|
|
__ lw(t2, MemOperand(fp, offsetof(T, y)) );
|
|
__ pop(t2);
|
|
__ nop();
|
|
__ push(t2);
|
|
__ lw(t2, MemOperand(fp, offsetof(T, y)) );
|
|
__ pop(t1);
|
|
__ nop();
|
|
__ push(t1);
|
|
__ lw(t2, MemOperand(fp, offsetof(T, y)) );
|
|
__ pop(t3);
|
|
__ nop();
|
|
|
|
__ mov(fp, t6);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
t.x = 1;
|
|
t.y = 2;
|
|
t.y1 = 3;
|
|
t.y2 = 4;
|
|
t.y3 = 0XBABA;
|
|
t.y4 = 0xDEDA;
|
|
|
|
f.Call(&t, 0, 0, 0, 0);
|
|
|
|
CHECK_EQ(3, t.y1);
|
|
}
|
|
|
|
|
|
TEST(MIPS13) {
|
|
// Test Cvt_d_uw and Trunc_uw_d macros.
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
struct T {
|
|
double cvt_big_out;
|
|
double cvt_small_out;
|
|
uint32_t trunc_big_out;
|
|
uint32_t trunc_small_out;
|
|
uint32_t cvt_big_in;
|
|
uint32_t cvt_small_in;
|
|
};
|
|
T t;
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
__ sw(t0, MemOperand(a0, offsetof(T, cvt_small_in)));
|
|
__ Cvt_d_uw(f10, t0, f4);
|
|
__ Sdc1(f10, MemOperand(a0, offsetof(T, cvt_small_out)));
|
|
|
|
__ Trunc_uw_d(f10, f10, f4);
|
|
__ swc1(f10, MemOperand(a0, offsetof(T, trunc_small_out)));
|
|
|
|
__ sw(t0, MemOperand(a0, offsetof(T, cvt_big_in)));
|
|
__ Cvt_d_uw(f8, t0, f4);
|
|
__ Sdc1(f8, MemOperand(a0, offsetof(T, cvt_big_out)));
|
|
|
|
__ Trunc_uw_d(f8, f8, f4);
|
|
__ swc1(f8, MemOperand(a0, offsetof(T, trunc_big_out)));
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
|
|
t.cvt_big_in = 0xFFFFFFFF;
|
|
t.cvt_small_in = 333;
|
|
|
|
f.Call(&t, 0, 0, 0, 0);
|
|
|
|
CHECK_EQ(t.cvt_big_out, static_cast<double>(t.cvt_big_in));
|
|
CHECK_EQ(t.cvt_small_out, static_cast<double>(t.cvt_small_in));
|
|
|
|
CHECK_EQ(static_cast<int>(t.trunc_big_out), static_cast<int>(t.cvt_big_in));
|
|
CHECK_EQ(static_cast<int>(t.trunc_small_out),
|
|
static_cast<int>(t.cvt_small_in));
|
|
}
|
|
|
|
|
|
TEST(MIPS14) {
|
|
// Test round, floor, ceil, trunc, cvt.
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
#define ROUND_STRUCT_ELEMENT(x) \
|
|
uint32_t x##_isNaN2008; \
|
|
int32_t x##_up_out; \
|
|
int32_t x##_down_out; \
|
|
int32_t neg_##x##_up_out; \
|
|
int32_t neg_##x##_down_out; \
|
|
uint32_t x##_err1_out; \
|
|
uint32_t x##_err2_out; \
|
|
uint32_t x##_err3_out; \
|
|
uint32_t x##_err4_out; \
|
|
int32_t x##_invalid_result;
|
|
|
|
struct T {
|
|
double round_up_in;
|
|
double round_down_in;
|
|
double neg_round_up_in;
|
|
double neg_round_down_in;
|
|
double err1_in;
|
|
double err2_in;
|
|
double err3_in;
|
|
double err4_in;
|
|
|
|
ROUND_STRUCT_ELEMENT(round)
|
|
ROUND_STRUCT_ELEMENT(floor)
|
|
ROUND_STRUCT_ELEMENT(ceil)
|
|
ROUND_STRUCT_ELEMENT(trunc)
|
|
ROUND_STRUCT_ELEMENT(cvt)
|
|
};
|
|
T t;
|
|
|
|
#undef ROUND_STRUCT_ELEMENT
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
// Save FCSR.
|
|
__ cfc1(a1, FCSR);
|
|
// Disable FPU exceptions.
|
|
__ ctc1(zero_reg, FCSR);
|
|
#define RUN_ROUND_TEST(x) \
|
|
__ cfc1(t0, FCSR); \
|
|
__ sw(t0, MemOperand(a0, offsetof(T, x##_isNaN2008))); \
|
|
__ Ldc1(f0, MemOperand(a0, offsetof(T, round_up_in))); \
|
|
__ x##_w_d(f0, f0); \
|
|
__ swc1(f0, MemOperand(a0, offsetof(T, x##_up_out))); \
|
|
\
|
|
__ Ldc1(f0, MemOperand(a0, offsetof(T, round_down_in))); \
|
|
__ x##_w_d(f0, f0); \
|
|
__ swc1(f0, MemOperand(a0, offsetof(T, x##_down_out))); \
|
|
\
|
|
__ Ldc1(f0, MemOperand(a0, offsetof(T, neg_round_up_in))); \
|
|
__ x##_w_d(f0, f0); \
|
|
__ swc1(f0, MemOperand(a0, offsetof(T, neg_##x##_up_out))); \
|
|
\
|
|
__ Ldc1(f0, MemOperand(a0, offsetof(T, neg_round_down_in))); \
|
|
__ x##_w_d(f0, f0); \
|
|
__ swc1(f0, MemOperand(a0, offsetof(T, neg_##x##_down_out))); \
|
|
\
|
|
__ Ldc1(f0, MemOperand(a0, offsetof(T, err1_in))); \
|
|
__ ctc1(zero_reg, FCSR); \
|
|
__ x##_w_d(f0, f0); \
|
|
__ cfc1(a2, FCSR); \
|
|
__ sw(a2, MemOperand(a0, offsetof(T, x##_err1_out))); \
|
|
\
|
|
__ Ldc1(f0, MemOperand(a0, offsetof(T, err2_in))); \
|
|
__ ctc1(zero_reg, FCSR); \
|
|
__ x##_w_d(f0, f0); \
|
|
__ cfc1(a2, FCSR); \
|
|
__ sw(a2, MemOperand(a0, offsetof(T, x##_err2_out))); \
|
|
\
|
|
__ Ldc1(f0, MemOperand(a0, offsetof(T, err3_in))); \
|
|
__ ctc1(zero_reg, FCSR); \
|
|
__ x##_w_d(f0, f0); \
|
|
__ cfc1(a2, FCSR); \
|
|
__ sw(a2, MemOperand(a0, offsetof(T, x##_err3_out))); \
|
|
\
|
|
__ Ldc1(f0, MemOperand(a0, offsetof(T, err4_in))); \
|
|
__ ctc1(zero_reg, FCSR); \
|
|
__ x##_w_d(f0, f0); \
|
|
__ cfc1(a2, FCSR); \
|
|
__ sw(a2, MemOperand(a0, offsetof(T, x##_err4_out))); \
|
|
__ swc1(f0, MemOperand(a0, offsetof(T, x##_invalid_result)));
|
|
|
|
RUN_ROUND_TEST(round)
|
|
RUN_ROUND_TEST(floor)
|
|
RUN_ROUND_TEST(ceil)
|
|
RUN_ROUND_TEST(trunc)
|
|
RUN_ROUND_TEST(cvt)
|
|
|
|
// Restore FCSR.
|
|
__ ctc1(a1, FCSR);
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
|
|
t.round_up_in = 123.51;
|
|
t.round_down_in = 123.49;
|
|
t.neg_round_up_in = -123.5;
|
|
t.neg_round_down_in = -123.49;
|
|
t.err1_in = 123.51;
|
|
t.err2_in = 1;
|
|
t.err3_in = static_cast<double>(1) + 0xFFFFFFFF;
|
|
t.err4_in = NAN;
|
|
|
|
f.Call(&t, 0, 0, 0, 0);
|
|
|
|
#define GET_FPU_ERR(x) (static_cast<int>(x & kFCSRFlagMask))
|
|
#define CHECK_NAN2008(x) (x & kFCSRNaN2008FlagMask)
|
|
#define CHECK_ROUND_RESULT(type) \
|
|
CHECK(GET_FPU_ERR(t.type##_err1_out) & kFCSRInexactFlagMask); \
|
|
CHECK_EQ(0, GET_FPU_ERR(t.type##_err2_out)); \
|
|
CHECK(GET_FPU_ERR(t.type##_err3_out) & kFCSRInvalidOpFlagMask); \
|
|
CHECK(GET_FPU_ERR(t.type##_err4_out) & kFCSRInvalidOpFlagMask); \
|
|
if (CHECK_NAN2008(t.type##_isNaN2008) && kArchVariant == kMips32r6) {\
|
|
CHECK_EQ(static_cast<int32_t>(0), t.type##_invalid_result);\
|
|
} else {\
|
|
CHECK_EQ(static_cast<int32_t>(kFPUInvalidResult), t.type##_invalid_result);\
|
|
}
|
|
|
|
|
|
CHECK_ROUND_RESULT(round);
|
|
CHECK_ROUND_RESULT(floor);
|
|
CHECK_ROUND_RESULT(ceil);
|
|
CHECK_ROUND_RESULT(cvt);
|
|
}
|
|
|
|
|
|
TEST(MIPS15) {
|
|
// Test chaining of label usages within instructions (issue 1644).
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
Assembler assm(AssemblerOptions{});
|
|
|
|
Label target;
|
|
__ beq(v0, v1, &target);
|
|
__ nop();
|
|
__ bne(v0, v1, &target);
|
|
__ nop();
|
|
__ bind(&target);
|
|
__ nop();
|
|
}
|
|
|
|
|
|
// ----------------------mips32r6 specific tests----------------------
|
|
TEST(seleqz_selnez) {
|
|
if (IsMipsArchVariant(kMips32r6)) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
struct Test {
|
|
int a;
|
|
int b;
|
|
int c;
|
|
int d;
|
|
double e;
|
|
double f;
|
|
double g;
|
|
double h;
|
|
float i;
|
|
float j;
|
|
float k;
|
|
float l;
|
|
};
|
|
|
|
Test test;
|
|
// Integer part of test.
|
|
__ addiu(t1, zero_reg, 1); // t1 = 1
|
|
__ seleqz(t3, t1, zero_reg); // t3 = 1
|
|
__ sw(t3, MemOperand(a0, offsetof(Test, a))); // a = 1
|
|
__ seleqz(t2, t1, t1); // t2 = 0
|
|
__ sw(t2, MemOperand(a0, offsetof(Test, b))); // b = 0
|
|
__ selnez(t3, t1, zero_reg); // t3 = 1;
|
|
__ sw(t3, MemOperand(a0, offsetof(Test, c))); // c = 0
|
|
__ selnez(t3, t1, t1); // t3 = 1
|
|
__ sw(t3, MemOperand(a0, offsetof(Test, d))); // d = 1
|
|
// Floating point part of test.
|
|
__ Ldc1(f0, MemOperand(a0, offsetof(Test, e))); // src
|
|
__ Ldc1(f2, MemOperand(a0, offsetof(Test, f))); // test
|
|
__ lwc1(f8, MemOperand(a0, offsetof(Test, i)) ); // src
|
|
__ lwc1(f10, MemOperand(a0, offsetof(Test, j)) ); // test
|
|
__ seleqz_d(f4, f0, f2);
|
|
__ selnez_d(f6, f0, f2);
|
|
__ seleqz_s(f12, f8, f10);
|
|
__ selnez_s(f14, f8, f10);
|
|
__ Sdc1(f4, MemOperand(a0, offsetof(Test, g))); // src
|
|
__ Sdc1(f6, MemOperand(a0, offsetof(Test, h))); // src
|
|
__ swc1(f12, MemOperand(a0, offsetof(Test, k)) ); // src
|
|
__ swc1(f14, MemOperand(a0, offsetof(Test, l)) ); // src
|
|
__ jr(ra);
|
|
__ nop();
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
|
|
CHECK_EQ(1, test.a);
|
|
CHECK_EQ(0, test.b);
|
|
CHECK_EQ(0, test.c);
|
|
CHECK_EQ(1, test.d);
|
|
|
|
const int test_size = 3;
|
|
const int input_size = 5;
|
|
|
|
double inputs_D[input_size] = {0.0, 65.2, -70.32,
|
|
18446744073709551621.0, -18446744073709551621.0};
|
|
double outputs_D[input_size] = {0.0, 65.2, -70.32,
|
|
18446744073709551621.0, -18446744073709551621.0};
|
|
double tests_D[test_size*2] = {2.8, 2.9, -2.8, -2.9,
|
|
18446744073709551616.0, 18446744073709555712.0};
|
|
float inputs_S[input_size] = {0.0, 65.2, -70.32,
|
|
18446744073709551621.0, -18446744073709551621.0};
|
|
float outputs_S[input_size] = {0.0, 65.2, -70.32,
|
|
18446744073709551621.0, -18446744073709551621.0};
|
|
float tests_S[test_size*2] = {2.9, 2.8, -2.9, -2.8,
|
|
18446744073709551616.0, 18446746272732807168.0};
|
|
for (int j=0; j < test_size; j+=2) {
|
|
for (int i=0; i < input_size; i++) {
|
|
test.e = inputs_D[i];
|
|
test.f = tests_D[j];
|
|
test.i = inputs_S[i];
|
|
test.j = tests_S[j];
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(outputs_D[i], test.g);
|
|
CHECK_EQ(0, test.h);
|
|
CHECK_EQ(outputs_S[i], test.k);
|
|
CHECK_EQ(0, test.l);
|
|
|
|
test.f = tests_D[j+1];
|
|
test.j = tests_S[j+1];
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(0, test.g);
|
|
CHECK_EQ(outputs_D[i], test.h);
|
|
CHECK_EQ(0, test.k);
|
|
CHECK_EQ(outputs_S[i], test.l);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
TEST(min_max) {
|
|
if (IsMipsArchVariant(kMips32r6)) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
struct TestFloat {
|
|
double a;
|
|
double b;
|
|
double c;
|
|
double d;
|
|
float e;
|
|
float f;
|
|
float g;
|
|
float h;
|
|
};
|
|
|
|
TestFloat test;
|
|
const double dnan = std::numeric_limits<double>::quiet_NaN();
|
|
const double dinf = std::numeric_limits<double>::infinity();
|
|
const double dminf = -std::numeric_limits<double>::infinity();
|
|
const float fnan = std::numeric_limits<float>::quiet_NaN();
|
|
const float finf = std::numeric_limits<float>::infinity();
|
|
const float fminf = std::numeric_limits<float>::infinity();
|
|
const int kTableLength = 13;
|
|
double inputsa[kTableLength] = {2.0, 3.0, dnan, 3.0, -0.0, 0.0, dinf,
|
|
dnan, 42.0, dinf, dminf, dinf, dnan};
|
|
double inputsb[kTableLength] = {3.0, 2.0, 3.0, dnan, 0.0, -0.0, dnan,
|
|
dinf, dinf, 42.0, dinf, dminf, dnan};
|
|
double outputsdmin[kTableLength] = {2.0, 2.0, 3.0, 3.0, -0.0,
|
|
-0.0, dinf, dinf, 42.0, 42.0,
|
|
dminf, dminf, dnan};
|
|
double outputsdmax[kTableLength] = {3.0, 3.0, 3.0, 3.0, 0.0, 0.0, dinf,
|
|
dinf, dinf, dinf, dinf, dinf, dnan};
|
|
|
|
float inputse[kTableLength] = {2.0, 3.0, fnan, 3.0, -0.0, 0.0, finf,
|
|
fnan, 42.0, finf, fminf, finf, fnan};
|
|
float inputsf[kTableLength] = {3.0, 2.0, 3.0, fnan, 0.0, -0.0, fnan,
|
|
finf, finf, 42.0, finf, fminf, fnan};
|
|
float outputsfmin[kTableLength] = {2.0, 2.0, 3.0, 3.0, -0.0,
|
|
-0.0, finf, finf, 42.0, 42.0,
|
|
fminf, fminf, fnan};
|
|
float outputsfmax[kTableLength] = {3.0, 3.0, 3.0, 3.0, 0.0, 0.0, finf,
|
|
finf, finf, finf, finf, finf, fnan};
|
|
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(TestFloat, a)));
|
|
__ Ldc1(f8, MemOperand(a0, offsetof(TestFloat, b)));
|
|
__ lwc1(f2, MemOperand(a0, offsetof(TestFloat, e)));
|
|
__ lwc1(f6, MemOperand(a0, offsetof(TestFloat, f)));
|
|
__ min_d(f10, f4, f8);
|
|
__ max_d(f12, f4, f8);
|
|
__ min_s(f14, f2, f6);
|
|
__ max_s(f16, f2, f6);
|
|
__ Sdc1(f10, MemOperand(a0, offsetof(TestFloat, c)));
|
|
__ Sdc1(f12, MemOperand(a0, offsetof(TestFloat, d)));
|
|
__ swc1(f14, MemOperand(a0, offsetof(TestFloat, g)));
|
|
__ swc1(f16, MemOperand(a0, offsetof(TestFloat, h)));
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
for (int i = 0; i < kTableLength; i++) {
|
|
test.a = inputsa[i];
|
|
test.b = inputsb[i];
|
|
test.e = inputse[i];
|
|
test.f = inputsf[i];
|
|
|
|
f.Call(&test, 0, 0, 0, 0);
|
|
|
|
CHECK_EQ(0, memcmp(&test.c, &outputsdmin[i], sizeof(test.c)));
|
|
CHECK_EQ(0, memcmp(&test.d, &outputsdmax[i], sizeof(test.d)));
|
|
CHECK_EQ(0, memcmp(&test.g, &outputsfmin[i], sizeof(test.g)));
|
|
CHECK_EQ(0, memcmp(&test.h, &outputsfmax[i], sizeof(test.h)));
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
TEST(rint_d) {
|
|
if (IsMipsArchVariant(kMips32r6)) {
|
|
const int kTableLength = 30;
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
struct TestFloat {
|
|
double a;
|
|
double b;
|
|
int fcsr;
|
|
};
|
|
|
|
TestFloat test;
|
|
double inputs[kTableLength] = {18446744073709551617.0,
|
|
4503599627370496.0, -4503599627370496.0,
|
|
1.26782468584154733584017312973E30, 1.44860108245951772690707170478E147,
|
|
1.7976931348623157E+308, 6.27463370218383111104242366943E-307,
|
|
309485009821345068724781056.89,
|
|
2.1, 2.6, 2.5, 3.1, 3.6, 3.5,
|
|
-2.1, -2.6, -2.5, -3.1, -3.6, -3.5,
|
|
37778931862957161709568.0, 37778931862957161709569.0,
|
|
37778931862957161709580.0, 37778931862957161709581.0,
|
|
37778931862957161709582.0, 37778931862957161709583.0,
|
|
37778931862957161709584.0, 37778931862957161709585.0,
|
|
37778931862957161709586.0, 37778931862957161709587.0};
|
|
double outputs_RN[kTableLength] = {18446744073709551617.0,
|
|
4503599627370496.0, -4503599627370496.0,
|
|
1.26782468584154733584017312973E30, 1.44860108245951772690707170478E147,
|
|
1.7976931348623157E308, 0,
|
|
309485009821345068724781057.0,
|
|
2.0, 3.0, 2.0, 3.0, 4.0, 4.0,
|
|
-2.0, -3.0, -2.0, -3.0, -4.0, -4.0,
|
|
37778931862957161709568.0, 37778931862957161709569.0,
|
|
37778931862957161709580.0, 37778931862957161709581.0,
|
|
37778931862957161709582.0, 37778931862957161709583.0,
|
|
37778931862957161709584.0, 37778931862957161709585.0,
|
|
37778931862957161709586.0, 37778931862957161709587.0};
|
|
double outputs_RZ[kTableLength] = {18446744073709551617.0,
|
|
4503599627370496.0, -4503599627370496.0,
|
|
1.26782468584154733584017312973E30, 1.44860108245951772690707170478E147,
|
|
1.7976931348623157E308, 0,
|
|
309485009821345068724781057.0,
|
|
2.0, 2.0, 2.0, 3.0, 3.0, 3.0,
|
|
-2.0, -2.0, -2.0, -3.0, -3.0, -3.0,
|
|
37778931862957161709568.0, 37778931862957161709569.0,
|
|
37778931862957161709580.0, 37778931862957161709581.0,
|
|
37778931862957161709582.0, 37778931862957161709583.0,
|
|
37778931862957161709584.0, 37778931862957161709585.0,
|
|
37778931862957161709586.0, 37778931862957161709587.0};
|
|
double outputs_RP[kTableLength] = {18446744073709551617.0,
|
|
4503599627370496.0, -4503599627370496.0,
|
|
1.26782468584154733584017312973E30, 1.44860108245951772690707170478E147,
|
|
1.7976931348623157E308, 1,
|
|
309485009821345068724781057.0,
|
|
3.0, 3.0, 3.0, 4.0, 4.0, 4.0,
|
|
-2.0, -2.0, -2.0, -3.0, -3.0, -3.0,
|
|
37778931862957161709568.0, 37778931862957161709569.0,
|
|
37778931862957161709580.0, 37778931862957161709581.0,
|
|
37778931862957161709582.0, 37778931862957161709583.0,
|
|
37778931862957161709584.0, 37778931862957161709585.0,
|
|
37778931862957161709586.0, 37778931862957161709587.0};
|
|
double outputs_RM[kTableLength] = {18446744073709551617.0,
|
|
4503599627370496.0, -4503599627370496.0,
|
|
1.26782468584154733584017312973E30, 1.44860108245951772690707170478E147,
|
|
1.7976931348623157E308, 0,
|
|
309485009821345068724781057.0,
|
|
2.0, 2.0, 2.0, 3.0, 3.0, 3.0,
|
|
-3.0, -3.0, -3.0, -4.0, -4.0, -4.0,
|
|
37778931862957161709568.0, 37778931862957161709569.0,
|
|
37778931862957161709580.0, 37778931862957161709581.0,
|
|
37778931862957161709582.0, 37778931862957161709583.0,
|
|
37778931862957161709584.0, 37778931862957161709585.0,
|
|
37778931862957161709586.0, 37778931862957161709587.0};
|
|
int fcsr_inputs[4] =
|
|
{kRoundToNearest, kRoundToZero, kRoundToPlusInf, kRoundToMinusInf};
|
|
double* outputs[4] = {outputs_RN, outputs_RZ, outputs_RP, outputs_RM};
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(TestFloat, a)));
|
|
__ lw(t0, MemOperand(a0, offsetof(TestFloat, fcsr)) );
|
|
__ cfc1(t1, FCSR);
|
|
__ ctc1(t0, FCSR);
|
|
__ rint_d(f8, f4);
|
|
__ Sdc1(f8, MemOperand(a0, offsetof(TestFloat, b)));
|
|
__ ctc1(t1, FCSR);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
|
|
for (int j = 0; j < 4; j++) {
|
|
test.fcsr = fcsr_inputs[j];
|
|
for (int i = 0; i < kTableLength; i++) {
|
|
test.a = inputs[i];
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.b, outputs[j][i]);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
TEST(sel) {
|
|
if (IsMipsArchVariant(kMips32r6)) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
struct Test {
|
|
double dd;
|
|
double ds;
|
|
double dt;
|
|
float fd;
|
|
float fs;
|
|
float ft;
|
|
};
|
|
|
|
Test test;
|
|
__ Ldc1(f0, MemOperand(a0, offsetof(Test, dd))); // test
|
|
__ Ldc1(f2, MemOperand(a0, offsetof(Test, ds))); // src1
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(Test, dt))); // src2
|
|
__ lwc1(f6, MemOperand(a0, offsetof(Test, fd)) ); // test
|
|
__ lwc1(f8, MemOperand(a0, offsetof(Test, fs)) ); // src1
|
|
__ lwc1(f10, MemOperand(a0, offsetof(Test, ft)) ); // src2
|
|
__ sel_d(f0, f2, f4);
|
|
__ sel_s(f6, f8, f10);
|
|
__ Sdc1(f0, MemOperand(a0, offsetof(Test, dd)));
|
|
__ swc1(f6, MemOperand(a0, offsetof(Test, fd)) );
|
|
__ jr(ra);
|
|
__ nop();
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
|
|
const int test_size = 3;
|
|
const int input_size = 5;
|
|
|
|
double inputs_dt[input_size] = {0.0, 65.2, -70.32,
|
|
18446744073709551621.0, -18446744073709551621.0};
|
|
double inputs_ds[input_size] = {0.1, 69.88, -91.325,
|
|
18446744073709551625.0, -18446744073709551625.0};
|
|
float inputs_ft[input_size] = {0.0, 65.2, -70.32,
|
|
18446744073709551621.0, -18446744073709551621.0};
|
|
float inputs_fs[input_size] = {0.1, 69.88, -91.325,
|
|
18446744073709551625.0, -18446744073709551625.0};
|
|
double tests_D[test_size*2] = {2.8, 2.9, -2.8, -2.9,
|
|
18446744073709551616.0, 18446744073709555712.0};
|
|
float tests_S[test_size*2] = {2.9, 2.8, -2.9, -2.8,
|
|
18446744073709551616.0, 18446746272732807168.0};
|
|
for (int j=0; j < test_size; j+=2) {
|
|
for (int i=0; i < input_size; i++) {
|
|
test.dt = inputs_dt[i];
|
|
test.dd = tests_D[j];
|
|
test.ds = inputs_ds[i];
|
|
test.ft = inputs_ft[i];
|
|
test.fd = tests_S[j];
|
|
test.fs = inputs_fs[i];
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.dd, inputs_ds[i]);
|
|
CHECK_EQ(test.fd, inputs_fs[i]);
|
|
|
|
test.dd = tests_D[j+1];
|
|
test.fd = tests_S[j+1];
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.dd, inputs_dt[i]);
|
|
CHECK_EQ(test.fd, inputs_ft[i]);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
TEST(rint_s) {
|
|
if (IsMipsArchVariant(kMips32r6)) {
|
|
const int kTableLength = 30;
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
struct TestFloat {
|
|
float a;
|
|
float b;
|
|
int fcsr;
|
|
};
|
|
|
|
TestFloat test;
|
|
float inputs[kTableLength] = {18446744073709551617.0,
|
|
4503599627370496.0, -4503599627370496.0,
|
|
1.26782468584154733584017312973E30, 1.44860108245951772690707170478E37,
|
|
1.7976931348623157E+38, 6.27463370218383111104242366943E-37,
|
|
309485009821345068724781056.89,
|
|
2.1, 2.6, 2.5, 3.1, 3.6, 3.5,
|
|
-2.1, -2.6, -2.5, -3.1, -3.6, -3.5,
|
|
37778931862957161709568.0, 37778931862957161709569.0,
|
|
37778931862957161709580.0, 37778931862957161709581.0,
|
|
37778931862957161709582.0, 37778931862957161709583.0,
|
|
37778931862957161709584.0, 37778931862957161709585.0,
|
|
37778931862957161709586.0, 37778931862957161709587.0};
|
|
float outputs_RN[kTableLength] = {18446744073709551617.0,
|
|
4503599627370496.0, -4503599627370496.0,
|
|
1.26782468584154733584017312973E30, 1.44860108245951772690707170478E37,
|
|
1.7976931348623157E38, 0,
|
|
309485009821345068724781057.0,
|
|
2.0, 3.0, 2.0, 3.0, 4.0, 4.0,
|
|
-2.0, -3.0, -2.0, -3.0, -4.0, -4.0,
|
|
37778931862957161709568.0, 37778931862957161709569.0,
|
|
37778931862957161709580.0, 37778931862957161709581.0,
|
|
37778931862957161709582.0, 37778931862957161709583.0,
|
|
37778931862957161709584.0, 37778931862957161709585.0,
|
|
37778931862957161709586.0, 37778931862957161709587.0};
|
|
float outputs_RZ[kTableLength] = {18446744073709551617.0,
|
|
4503599627370496.0, -4503599627370496.0,
|
|
1.26782468584154733584017312973E30, 1.44860108245951772690707170478E37,
|
|
1.7976931348623157E38, 0,
|
|
309485009821345068724781057.0,
|
|
2.0, 2.0, 2.0, 3.0, 3.0, 3.0,
|
|
-2.0, -2.0, -2.0, -3.0, -3.0, -3.0,
|
|
37778931862957161709568.0, 37778931862957161709569.0,
|
|
37778931862957161709580.0, 37778931862957161709581.0,
|
|
37778931862957161709582.0, 37778931862957161709583.0,
|
|
37778931862957161709584.0, 37778931862957161709585.0,
|
|
37778931862957161709586.0, 37778931862957161709587.0};
|
|
float outputs_RP[kTableLength] = {18446744073709551617.0,
|
|
4503599627370496.0, -4503599627370496.0,
|
|
1.26782468584154733584017312973E30, 1.44860108245951772690707170478E37,
|
|
1.7976931348623157E38, 1,
|
|
309485009821345068724781057.0,
|
|
3.0, 3.0, 3.0, 4.0, 4.0, 4.0,
|
|
-2.0, -2.0, -2.0, -3.0, -3.0, -3.0,
|
|
37778931862957161709568.0, 37778931862957161709569.0,
|
|
37778931862957161709580.0, 37778931862957161709581.0,
|
|
37778931862957161709582.0, 37778931862957161709583.0,
|
|
37778931862957161709584.0, 37778931862957161709585.0,
|
|
37778931862957161709586.0, 37778931862957161709587.0};
|
|
float outputs_RM[kTableLength] = {18446744073709551617.0,
|
|
4503599627370496.0, -4503599627370496.0,
|
|
1.26782468584154733584017312973E30, 1.44860108245951772690707170478E37,
|
|
1.7976931348623157E38, 0,
|
|
309485009821345068724781057.0,
|
|
2.0, 2.0, 2.0, 3.0, 3.0, 3.0,
|
|
-3.0, -3.0, -3.0, -4.0, -4.0, -4.0,
|
|
37778931862957161709568.0, 37778931862957161709569.0,
|
|
37778931862957161709580.0, 37778931862957161709581.0,
|
|
37778931862957161709582.0, 37778931862957161709583.0,
|
|
37778931862957161709584.0, 37778931862957161709585.0,
|
|
37778931862957161709586.0, 37778931862957161709587.0};
|
|
int fcsr_inputs[4] =
|
|
{kRoundToNearest, kRoundToZero, kRoundToPlusInf, kRoundToMinusInf};
|
|
float* outputs[4] = {outputs_RN, outputs_RZ, outputs_RP, outputs_RM};
|
|
__ lwc1(f4, MemOperand(a0, offsetof(TestFloat, a)) );
|
|
__ lw(t0, MemOperand(a0, offsetof(TestFloat, fcsr)) );
|
|
__ cfc1(t1, FCSR);
|
|
__ ctc1(t0, FCSR);
|
|
__ rint_s(f8, f4);
|
|
__ swc1(f8, MemOperand(a0, offsetof(TestFloat, b)) );
|
|
__ ctc1(t1, FCSR);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
|
|
for (int j = 0; j < 4; j++) {
|
|
test.fcsr = fcsr_inputs[j];
|
|
for (int i = 0; i < kTableLength; i++) {
|
|
test.a = inputs[i];
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.b, outputs[j][i]);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
TEST(Cvt_d_uw) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
struct TestStruct {
|
|
unsigned input;
|
|
uint64_t output;
|
|
};
|
|
|
|
unsigned inputs[] = {0x0, 0xFFFFFFFF, 0x80000000, 0x7FFFFFFF};
|
|
|
|
uint64_t outputs[] = {0x0, 0x41EFFFFFFFE00000, 0x41E0000000000000,
|
|
0x41DFFFFFFFC00000};
|
|
|
|
int kTableLength = sizeof(inputs)/sizeof(inputs[0]);
|
|
|
|
TestStruct test;
|
|
|
|
__ lw(t1, MemOperand(a0, offsetof(TestStruct, input)));
|
|
__ Cvt_d_uw(f4, t1, f6);
|
|
__ Sdc1(f4, MemOperand(a0, offsetof(TestStruct, output)));
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
for (int i = 0; i < kTableLength; i++) {
|
|
test.input = inputs[i];
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
// Check outputs
|
|
CHECK_EQ(test.output, outputs[i]);
|
|
}
|
|
}
|
|
|
|
|
|
TEST(mina_maxa) {
|
|
if (IsMipsArchVariant(kMips32r6)) {
|
|
const int kTableLength = 23;
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
const double dnan = std::numeric_limits<double>::quiet_NaN();
|
|
const double dinf = std::numeric_limits<double>::infinity();
|
|
const double dminf = -std::numeric_limits<double>::infinity();
|
|
const float fnan = std::numeric_limits<float>::quiet_NaN();
|
|
const float finf = std::numeric_limits<float>::infinity();
|
|
const float fminf = std::numeric_limits<float>::infinity();
|
|
|
|
struct TestFloat {
|
|
double a;
|
|
double b;
|
|
double resd;
|
|
double resd1;
|
|
float c;
|
|
float d;
|
|
float resf;
|
|
float resf1;
|
|
};
|
|
|
|
TestFloat test;
|
|
double inputsa[kTableLength] = {
|
|
5.3, 4.8, 6.1, 9.8, 9.8, 9.8, -10.0, -8.9, -9.8, -10.0, -8.9, -9.8,
|
|
dnan, 3.0, -0.0, 0.0, dinf, dnan, 42.0, dinf, dminf, dinf, dnan};
|
|
double inputsb[kTableLength] = {
|
|
4.8, 5.3, 6.1, -10.0, -8.9, -9.8, 9.8, 9.8, 9.8, -9.8, -11.2, -9.8,
|
|
3.0, dnan, 0.0, -0.0, dnan, dinf, dinf, 42.0, dinf, dminf, dnan};
|
|
double resd[kTableLength] = {
|
|
4.8, 4.8, 6.1, 9.8, -8.9, -9.8, 9.8, -8.9, -9.8, -9.8, -8.9, -9.8,
|
|
3.0, 3.0, -0.0, -0.0, dinf, dinf, 42.0, 42.0, dminf, dminf, dnan};
|
|
double resd1[kTableLength] = {
|
|
5.3, 5.3, 6.1, -10.0, 9.8, 9.8, -10.0, 9.8, 9.8, -10.0, -11.2, -9.8,
|
|
3.0, 3.0, 0.0, 0.0, dinf, dinf, dinf, dinf, dinf, dinf, dnan};
|
|
float inputsc[kTableLength] = {
|
|
5.3, 4.8, 6.1, 9.8, 9.8, 9.8, -10.0, -8.9, -9.8, -10.0, -8.9, -9.8,
|
|
fnan, 3.0, -0.0, 0.0, finf, fnan, 42.0, finf, fminf, finf, fnan};
|
|
float inputsd[kTableLength] = {4.8, 5.3, 6.1, -10.0, -8.9, -9.8,
|
|
9.8, 9.8, 9.8, -9.8, -11.2, -9.8,
|
|
3.0, fnan, -0.0, 0.0, fnan, finf,
|
|
finf, 42.0, finf, fminf, fnan};
|
|
float resf[kTableLength] = {
|
|
4.8, 4.8, 6.1, 9.8, -8.9, -9.8, 9.8, -8.9, -9.8, -9.8, -8.9, -9.8,
|
|
3.0, 3.0, -0.0, -0.0, finf, finf, 42.0, 42.0, fminf, fminf, fnan};
|
|
float resf1[kTableLength] = {
|
|
5.3, 5.3, 6.1, -10.0, 9.8, 9.8, -10.0, 9.8, 9.8, -10.0, -11.2, -9.8,
|
|
3.0, 3.0, 0.0, 0.0, finf, finf, finf, finf, finf, finf, fnan};
|
|
|
|
__ Ldc1(f2, MemOperand(a0, offsetof(TestFloat, a)));
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(TestFloat, b)));
|
|
__ lwc1(f8, MemOperand(a0, offsetof(TestFloat, c)) );
|
|
__ lwc1(f10, MemOperand(a0, offsetof(TestFloat, d)) );
|
|
__ mina_d(f6, f2, f4);
|
|
__ mina_s(f12, f8, f10);
|
|
__ maxa_d(f14, f2, f4);
|
|
__ maxa_s(f16, f8, f10);
|
|
__ swc1(f12, MemOperand(a0, offsetof(TestFloat, resf)) );
|
|
__ Sdc1(f6, MemOperand(a0, offsetof(TestFloat, resd)));
|
|
__ swc1(f16, MemOperand(a0, offsetof(TestFloat, resf1)) );
|
|
__ Sdc1(f14, MemOperand(a0, offsetof(TestFloat, resd1)));
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
for (int i = 0; i < kTableLength; i++) {
|
|
test.a = inputsa[i];
|
|
test.b = inputsb[i];
|
|
test.c = inputsc[i];
|
|
test.d = inputsd[i];
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
if (i < kTableLength - 1) {
|
|
CHECK_EQ(test.resd, resd[i]);
|
|
CHECK_EQ(test.resf, resf[i]);
|
|
CHECK_EQ(test.resd1, resd1[i]);
|
|
CHECK_EQ(test.resf1, resf1[i]);
|
|
} else {
|
|
CHECK(std::isnan(test.resd));
|
|
CHECK(std::isnan(test.resf));
|
|
CHECK(std::isnan(test.resd1));
|
|
CHECK(std::isnan(test.resf1));
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
// ----------------------mips32r2 specific tests----------------------
|
|
TEST(trunc_l) {
|
|
if (IsMipsArchVariant(kMips32r2) && IsFp64Mode()) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
const double dFPU64InvalidResult = static_cast<double>(kFPU64InvalidResult);
|
|
struct Test {
|
|
uint32_t isNaN2008;
|
|
double a;
|
|
float b;
|
|
int64_t c; // a trunc result
|
|
int64_t d; // b trunc result
|
|
};
|
|
const int kTableLength = 15;
|
|
double inputs_D[kTableLength] = {
|
|
2.1, 2.6, 2.5, 3.1, 3.6, 3.5,
|
|
-2.1, -2.6, -2.5, -3.1, -3.6, -3.5,
|
|
2147483648.0,
|
|
std::numeric_limits<double>::quiet_NaN(),
|
|
std::numeric_limits<double>::infinity()
|
|
};
|
|
float inputs_S[kTableLength] = {
|
|
2.1, 2.6, 2.5, 3.1, 3.6, 3.5,
|
|
-2.1, -2.6, -2.5, -3.1, -3.6, -3.5,
|
|
2147483648.0,
|
|
std::numeric_limits<float>::quiet_NaN(),
|
|
std::numeric_limits<float>::infinity()
|
|
};
|
|
double outputs[kTableLength] = {
|
|
2.0, 2.0, 2.0, 3.0, 3.0, 3.0,
|
|
-2.0, -2.0, -2.0, -3.0, -3.0, -3.0,
|
|
2147483648.0, dFPU64InvalidResult,
|
|
dFPU64InvalidResult};
|
|
double outputsNaN2008[kTableLength] = {
|
|
2.0, 2.0, 2.0, 3.0, 3.0, 3.0,
|
|
-2.0, -2.0, -2.0, -3.0, -3.0, -3.0,
|
|
2147483648.0,
|
|
0,
|
|
dFPU64InvalidResult};
|
|
|
|
__ cfc1(t1, FCSR);
|
|
__ sw(t1, MemOperand(a0, offsetof(Test, isNaN2008)));
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(Test, a)));
|
|
__ lwc1(f6, MemOperand(a0, offsetof(Test, b)) );
|
|
__ trunc_l_d(f8, f4);
|
|
__ trunc_l_s(f10, f6);
|
|
__ Sdc1(f8, MemOperand(a0, offsetof(Test, c)));
|
|
__ Sdc1(f10, MemOperand(a0, offsetof(Test, d)));
|
|
__ jr(ra);
|
|
__ nop();
|
|
Test test;
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
for (int i = 0; i < kTableLength; i++) {
|
|
test.a = inputs_D[i];
|
|
test.b = inputs_S[i];
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
if ((test.isNaN2008 & kFCSRNaN2008FlagMask) &&
|
|
kArchVariant == kMips32r6) {
|
|
CHECK_EQ(test.c, outputsNaN2008[i]);
|
|
} else {
|
|
CHECK_EQ(test.c, outputs[i]);
|
|
}
|
|
CHECK_EQ(test.d, test.c);
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
TEST(movz_movn) {
|
|
if (IsMipsArchVariant(kMips32r2)) {
|
|
const int kTableLength = 4;
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
struct TestFloat {
|
|
int32_t rt;
|
|
double a;
|
|
double b;
|
|
double bold;
|
|
double b1;
|
|
double bold1;
|
|
float c;
|
|
float d;
|
|
float dold;
|
|
float d1;
|
|
float dold1;
|
|
};
|
|
|
|
TestFloat test;
|
|
double inputs_D[kTableLength] = {
|
|
5.3, -5.3, 5.3, -2.9
|
|
};
|
|
double inputs_S[kTableLength] = {
|
|
4.8, 4.8, -4.8, -0.29
|
|
};
|
|
|
|
float outputs_S[kTableLength] = {
|
|
4.8, 4.8, -4.8, -0.29
|
|
};
|
|
double outputs_D[kTableLength] = {
|
|
5.3, -5.3, 5.3, -2.9
|
|
};
|
|
|
|
__ Ldc1(f2, MemOperand(a0, offsetof(TestFloat, a)));
|
|
__ lwc1(f6, MemOperand(a0, offsetof(TestFloat, c)) );
|
|
__ lw(t0, MemOperand(a0, offsetof(TestFloat, rt)) );
|
|
__ Move(f12, 0.0);
|
|
__ Move(f10, 0.0);
|
|
__ Move(f16, 0.0);
|
|
__ Move(f14, 0.0);
|
|
__ Sdc1(f12, MemOperand(a0, offsetof(TestFloat, bold)));
|
|
__ swc1(f10, MemOperand(a0, offsetof(TestFloat, dold)) );
|
|
__ Sdc1(f16, MemOperand(a0, offsetof(TestFloat, bold1)));
|
|
__ swc1(f14, MemOperand(a0, offsetof(TestFloat, dold1)) );
|
|
__ movz_s(f10, f6, t0);
|
|
__ movz_d(f12, f2, t0);
|
|
__ movn_s(f14, f6, t0);
|
|
__ movn_d(f16, f2, t0);
|
|
__ swc1(f10, MemOperand(a0, offsetof(TestFloat, d)) );
|
|
__ Sdc1(f12, MemOperand(a0, offsetof(TestFloat, b)));
|
|
__ swc1(f14, MemOperand(a0, offsetof(TestFloat, d1)) );
|
|
__ Sdc1(f16, MemOperand(a0, offsetof(TestFloat, b1)));
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
for (int i = 0; i < kTableLength; i++) {
|
|
test.a = inputs_D[i];
|
|
test.c = inputs_S[i];
|
|
|
|
test.rt = 1;
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.b, test.bold);
|
|
CHECK_EQ(test.d, test.dold);
|
|
CHECK_EQ(test.b1, outputs_D[i]);
|
|
CHECK_EQ(test.d1, outputs_S[i]);
|
|
|
|
test.rt = 0;
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.b, outputs_D[i]);
|
|
CHECK_EQ(test.d, outputs_S[i]);
|
|
CHECK_EQ(test.b1, test.bold1);
|
|
CHECK_EQ(test.d1, test.dold1);
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
TEST(movt_movd) {
|
|
if (IsMipsArchVariant(kMips32r2)) {
|
|
const int kTableLength = 4;
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
|
|
struct TestFloat {
|
|
double srcd;
|
|
double dstd;
|
|
double dstdold;
|
|
double dstd1;
|
|
double dstdold1;
|
|
float srcf;
|
|
float dstf;
|
|
float dstfold;
|
|
float dstf1;
|
|
float dstfold1;
|
|
int32_t cc;
|
|
int32_t fcsr;
|
|
};
|
|
|
|
TestFloat test;
|
|
double inputs_D[kTableLength] = {
|
|
5.3, -5.3, 20.8, -2.9
|
|
};
|
|
double inputs_S[kTableLength] = {
|
|
4.88, 4.8, -4.8, -0.29
|
|
};
|
|
|
|
float outputs_S[kTableLength] = {
|
|
4.88, 4.8, -4.8, -0.29
|
|
};
|
|
double outputs_D[kTableLength] = {
|
|
5.3, -5.3, 20.8, -2.9
|
|
};
|
|
int condition_flags[8] = {0, 1, 2, 3, 4, 5, 6, 7};
|
|
|
|
for (int i = 0; i < kTableLength; i++) {
|
|
test.srcd = inputs_D[i];
|
|
test.srcf = inputs_S[i];
|
|
|
|
for (int j = 0; j< 8; j++) {
|
|
test.cc = condition_flags[j];
|
|
if (test.cc == 0) {
|
|
test.fcsr = 1 << 23;
|
|
} else {
|
|
test.fcsr = 1 << (24+condition_flags[j]);
|
|
}
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
__ Ldc1(f2, MemOperand(a0, offsetof(TestFloat, srcd)));
|
|
__ lwc1(f4, MemOperand(a0, offsetof(TestFloat, srcf)) );
|
|
__ lw(t1, MemOperand(a0, offsetof(TestFloat, fcsr)) );
|
|
__ cfc1(t0, FCSR);
|
|
__ ctc1(t1, FCSR);
|
|
__ li(t2, 0x0);
|
|
__ mtc1(t2, f12);
|
|
__ mtc1(t2, f10);
|
|
__ Sdc1(f10, MemOperand(a0, offsetof(TestFloat, dstdold)));
|
|
__ swc1(f12, MemOperand(a0, offsetof(TestFloat, dstfold)) );
|
|
__ movt_s(f12, f4, test.cc);
|
|
__ movt_d(f10, f2, test.cc);
|
|
__ swc1(f12, MemOperand(a0, offsetof(TestFloat, dstf)) );
|
|
__ Sdc1(f10, MemOperand(a0, offsetof(TestFloat, dstd)));
|
|
__ Sdc1(f10, MemOperand(a0, offsetof(TestFloat, dstdold1)));
|
|
__ swc1(f12, MemOperand(a0, offsetof(TestFloat, dstfold1)) );
|
|
__ movf_s(f12, f4, test.cc);
|
|
__ movf_d(f10, f2, test.cc);
|
|
__ swc1(f12, MemOperand(a0, offsetof(TestFloat, dstf1)) );
|
|
__ Sdc1(f10, MemOperand(a0, offsetof(TestFloat, dstd1)));
|
|
__ ctc1(t0, FCSR);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.dstf, outputs_S[i]);
|
|
CHECK_EQ(test.dstd, outputs_D[i]);
|
|
CHECK_EQ(test.dstf1, test.dstfold1);
|
|
CHECK_EQ(test.dstd1, test.dstdold1);
|
|
test.fcsr = 0;
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.dstf, test.dstfold);
|
|
CHECK_EQ(test.dstd, test.dstdold);
|
|
CHECK_EQ(test.dstf1, outputs_S[i]);
|
|
CHECK_EQ(test.dstd1, outputs_D[i]);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
// ----------------------tests for all archs--------------------------
|
|
TEST(cvt_w_d) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
struct Test {
|
|
double a;
|
|
int32_t b;
|
|
int32_t fcsr;
|
|
};
|
|
const int kTableLength = 24;
|
|
double inputs[kTableLength] = {
|
|
2.1, 2.6, 2.5, 3.1, 3.6, 3.5,
|
|
-2.1, -2.6, -2.5, -3.1, -3.6, -3.5,
|
|
2147483637.0, 2147483638.0, 2147483639.0,
|
|
2147483640.0, 2147483641.0, 2147483642.0,
|
|
2147483643.0, 2147483644.0, 2147483645.0,
|
|
2147483646.0, 2147483647.0, 2147483653.0
|
|
};
|
|
double outputs_RN[kTableLength] = {
|
|
2.0, 3.0, 2.0, 3.0, 4.0, 4.0,
|
|
-2.0, -3.0, -2.0, -3.0, -4.0, -4.0,
|
|
2147483637.0, 2147483638.0, 2147483639.0,
|
|
2147483640.0, 2147483641.0, 2147483642.0,
|
|
2147483643.0, 2147483644.0, 2147483645.0,
|
|
2147483646.0, 2147483647.0, kFPUInvalidResult};
|
|
double outputs_RZ[kTableLength] = {
|
|
2.0, 2.0, 2.0, 3.0, 3.0, 3.0,
|
|
-2.0, -2.0, -2.0, -3.0, -3.0, -3.0,
|
|
2147483637.0, 2147483638.0, 2147483639.0,
|
|
2147483640.0, 2147483641.0, 2147483642.0,
|
|
2147483643.0, 2147483644.0, 2147483645.0,
|
|
2147483646.0, 2147483647.0, kFPUInvalidResult};
|
|
double outputs_RP[kTableLength] = {
|
|
3.0, 3.0, 3.0, 4.0, 4.0, 4.0,
|
|
-2.0, -2.0, -2.0, -3.0, -3.0, -3.0,
|
|
2147483637.0, 2147483638.0, 2147483639.0,
|
|
2147483640.0, 2147483641.0, 2147483642.0,
|
|
2147483643.0, 2147483644.0, 2147483645.0,
|
|
2147483646.0, 2147483647.0, kFPUInvalidResult};
|
|
double outputs_RM[kTableLength] = {
|
|
2.0, 2.0, 2.0, 3.0, 3.0, 3.0,
|
|
-3.0, -3.0, -3.0, -4.0, -4.0, -4.0,
|
|
2147483637.0, 2147483638.0, 2147483639.0,
|
|
2147483640.0, 2147483641.0, 2147483642.0,
|
|
2147483643.0, 2147483644.0, 2147483645.0,
|
|
2147483646.0, 2147483647.0, kFPUInvalidResult};
|
|
int fcsr_inputs[4] =
|
|
{kRoundToNearest, kRoundToZero, kRoundToPlusInf, kRoundToMinusInf};
|
|
double* outputs[4] = {outputs_RN, outputs_RZ, outputs_RP, outputs_RM};
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(Test, a)));
|
|
__ lw(t0, MemOperand(a0, offsetof(Test, fcsr)) );
|
|
__ cfc1(t1, FCSR);
|
|
__ ctc1(t0, FCSR);
|
|
__ cvt_w_d(f8, f4);
|
|
__ swc1(f8, MemOperand(a0, offsetof(Test, b)) );
|
|
__ ctc1(t1, FCSR);
|
|
__ jr(ra);
|
|
__ nop();
|
|
Test test;
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
for (int j = 0; j < 4; j++) {
|
|
test.fcsr = fcsr_inputs[j];
|
|
for (int i = 0; i < kTableLength; i++) {
|
|
test.a = inputs[i];
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.b, outputs[j][i]);
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
TEST(trunc_w) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
struct Test {
|
|
uint32_t isNaN2008;
|
|
double a;
|
|
float b;
|
|
int32_t c; // a trunc result
|
|
int32_t d; // b trunc result
|
|
};
|
|
const int kTableLength = 15;
|
|
double inputs_D[kTableLength] = {
|
|
2.1, 2.6, 2.5, 3.1, 3.6, 3.5,
|
|
-2.1, -2.6, -2.5, -3.1, -3.6, -3.5,
|
|
2147483648.0,
|
|
std::numeric_limits<double>::quiet_NaN(),
|
|
std::numeric_limits<double>::infinity()
|
|
};
|
|
float inputs_S[kTableLength] = {
|
|
2.1, 2.6, 2.5, 3.1, 3.6, 3.5,
|
|
-2.1, -2.6, -2.5, -3.1, -3.6, -3.5,
|
|
2147483648.0,
|
|
std::numeric_limits<float>::quiet_NaN(),
|
|
std::numeric_limits<float>::infinity()
|
|
};
|
|
double outputs[kTableLength] = {
|
|
2.0, 2.0, 2.0, 3.0, 3.0, 3.0,
|
|
-2.0, -2.0, -2.0, -3.0, -3.0, -3.0,
|
|
kFPUInvalidResult, kFPUInvalidResult,
|
|
kFPUInvalidResult};
|
|
double outputsNaN2008[kTableLength] = {
|
|
2.0, 2.0, 2.0, 3.0, 3.0, 3.0,
|
|
-2.0, -2.0, -2.0, -3.0, -3.0, -3.0,
|
|
kFPUInvalidResult,
|
|
0,
|
|
kFPUInvalidResult};
|
|
|
|
__ cfc1(t1, FCSR);
|
|
__ sw(t1, MemOperand(a0, offsetof(Test, isNaN2008)));
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(Test, a)));
|
|
__ lwc1(f6, MemOperand(a0, offsetof(Test, b)) );
|
|
__ trunc_w_d(f8, f4);
|
|
__ trunc_w_s(f10, f6);
|
|
__ swc1(f8, MemOperand(a0, offsetof(Test, c)) );
|
|
__ swc1(f10, MemOperand(a0, offsetof(Test, d)) );
|
|
__ jr(ra);
|
|
__ nop();
|
|
Test test;
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
for (int i = 0; i < kTableLength; i++) {
|
|
test.a = inputs_D[i];
|
|
test.b = inputs_S[i];
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
if ((test.isNaN2008 & kFCSRNaN2008FlagMask) && kArchVariant == kMips32r6) {
|
|
CHECK_EQ(test.c, outputsNaN2008[i]);
|
|
} else {
|
|
CHECK_EQ(test.c, outputs[i]);
|
|
}
|
|
CHECK_EQ(test.d, test.c);
|
|
}
|
|
}
|
|
|
|
|
|
TEST(round_w) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
struct Test {
|
|
uint32_t isNaN2008;
|
|
double a;
|
|
float b;
|
|
int32_t c; // a trunc result
|
|
int32_t d; // b trunc result
|
|
};
|
|
const int kTableLength = 15;
|
|
double inputs_D[kTableLength] = {
|
|
2.1, 2.6, 2.5, 3.1, 3.6, 3.5,
|
|
-2.1, -2.6, -2.5, -3.1, -3.6, -3.5,
|
|
2147483648.0,
|
|
std::numeric_limits<double>::quiet_NaN(),
|
|
std::numeric_limits<double>::infinity()
|
|
};
|
|
float inputs_S[kTableLength] = {
|
|
2.1, 2.6, 2.5, 3.1, 3.6, 3.5,
|
|
-2.1, -2.6, -2.5, -3.1, -3.6, -3.5,
|
|
2147483648.0,
|
|
std::numeric_limits<float>::quiet_NaN(),
|
|
std::numeric_limits<float>::infinity()
|
|
};
|
|
double outputs[kTableLength] = {
|
|
2.0, 3.0, 2.0, 3.0, 4.0, 4.0,
|
|
-2.0, -3.0, -2.0, -3.0, -4.0, -4.0,
|
|
kFPUInvalidResult, kFPUInvalidResult,
|
|
kFPUInvalidResult};
|
|
double outputsNaN2008[kTableLength] = {
|
|
2.0, 3.0, 2.0, 3.0, 4.0, 4.0,
|
|
-2.0, -3.0, -2.0, -3.0, -4.0, -4.0,
|
|
kFPUInvalidResult, 0,
|
|
kFPUInvalidResult};
|
|
|
|
__ cfc1(t1, FCSR);
|
|
__ sw(t1, MemOperand(a0, offsetof(Test, isNaN2008)));
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(Test, a)));
|
|
__ lwc1(f6, MemOperand(a0, offsetof(Test, b)) );
|
|
__ round_w_d(f8, f4);
|
|
__ round_w_s(f10, f6);
|
|
__ swc1(f8, MemOperand(a0, offsetof(Test, c)) );
|
|
__ swc1(f10, MemOperand(a0, offsetof(Test, d)) );
|
|
__ jr(ra);
|
|
__ nop();
|
|
Test test;
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
for (int i = 0; i < kTableLength; i++) {
|
|
test.a = inputs_D[i];
|
|
test.b = inputs_S[i];
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
if ((test.isNaN2008 & kFCSRNaN2008FlagMask) && kArchVariant == kMips32r6) {
|
|
CHECK_EQ(test.c, outputsNaN2008[i]);
|
|
} else {
|
|
CHECK_EQ(test.c, outputs[i]);
|
|
}
|
|
CHECK_EQ(test.d, test.c);
|
|
}
|
|
}
|
|
|
|
|
|
TEST(round_l) {
|
|
if (IsFp64Mode()) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
const double dFPU64InvalidResult = static_cast<double>(kFPU64InvalidResult);
|
|
struct Test {
|
|
uint32_t isNaN2008;
|
|
double a;
|
|
float b;
|
|
int64_t c;
|
|
int64_t d;
|
|
};
|
|
const int kTableLength = 15;
|
|
double inputs_D[kTableLength] = {
|
|
2.1, 2.6, 2.5, 3.1, 3.6, 3.5,
|
|
-2.1, -2.6, -2.5, -3.1, -3.6, -3.5,
|
|
2147483648.0,
|
|
std::numeric_limits<double>::quiet_NaN(),
|
|
std::numeric_limits<double>::infinity()
|
|
};
|
|
float inputs_S[kTableLength] = {
|
|
2.1, 2.6, 2.5, 3.1, 3.6, 3.5,
|
|
-2.1, -2.6, -2.5, -3.1, -3.6, -3.5,
|
|
2147483648.0,
|
|
std::numeric_limits<float>::quiet_NaN(),
|
|
std::numeric_limits<float>::infinity()
|
|
};
|
|
double outputs[kTableLength] = {
|
|
2.0, 3.0, 2.0, 3.0, 4.0, 4.0,
|
|
-2.0, -3.0, -2.0, -3.0, -4.0, -4.0,
|
|
2147483648.0, dFPU64InvalidResult,
|
|
dFPU64InvalidResult};
|
|
double outputsNaN2008[kTableLength] = {
|
|
2.0, 3.0, 2.0, 3.0, 4.0, 4.0,
|
|
-2.0, -3.0, -2.0, -3.0, -4.0, -4.0,
|
|
2147483648.0,
|
|
0,
|
|
dFPU64InvalidResult};
|
|
|
|
__ cfc1(t1, FCSR);
|
|
__ sw(t1, MemOperand(a0, offsetof(Test, isNaN2008)));
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(Test, a)));
|
|
__ lwc1(f6, MemOperand(a0, offsetof(Test, b)) );
|
|
__ round_l_d(f8, f4);
|
|
__ round_l_s(f10, f6);
|
|
__ Sdc1(f8, MemOperand(a0, offsetof(Test, c)));
|
|
__ Sdc1(f10, MemOperand(a0, offsetof(Test, d)));
|
|
__ jr(ra);
|
|
__ nop();
|
|
Test test;
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
for (int i = 0; i < kTableLength; i++) {
|
|
test.a = inputs_D[i];
|
|
test.b = inputs_S[i];
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
if ((test.isNaN2008 & kFCSRNaN2008FlagMask) &&
|
|
kArchVariant == kMips32r6) {
|
|
CHECK_EQ(test.c, outputsNaN2008[i]);
|
|
} else {
|
|
CHECK_EQ(test.c, outputs[i]);
|
|
}
|
|
CHECK_EQ(test.d, test.c);
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
TEST(sub) {
|
|
const int kTableLength = 12;
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
struct TestFloat {
|
|
float a;
|
|
float b;
|
|
float resultS;
|
|
double c;
|
|
double d;
|
|
double resultD;
|
|
};
|
|
|
|
TestFloat test;
|
|
double inputfs_D[kTableLength] = {
|
|
5.3, 4.8, 2.9, -5.3, -4.8, -2.9,
|
|
5.3, 4.8, 2.9, -5.3, -4.8, -2.9
|
|
};
|
|
double inputft_D[kTableLength] = {
|
|
4.8, 5.3, 2.9, 4.8, 5.3, 2.9,
|
|
-4.8, -5.3, -2.9, -4.8, -5.3, -2.9
|
|
};
|
|
double outputs_D[kTableLength] = {
|
|
0.5, -0.5, 0.0, -10.1, -10.1, -5.8,
|
|
10.1, 10.1, 5.8, -0.5, 0.5, 0.0
|
|
};
|
|
float inputfs_S[kTableLength] = {
|
|
5.3, 4.8, 2.9, -5.3, -4.8, -2.9,
|
|
5.3, 4.8, 2.9, -5.3, -4.8, -2.9
|
|
};
|
|
float inputft_S[kTableLength] = {
|
|
4.8, 5.3, 2.9, 4.8, 5.3, 2.9,
|
|
-4.8, -5.3, -2.9, -4.8, -5.3, -2.9
|
|
};
|
|
float outputs_S[kTableLength] = {
|
|
0.5, -0.5, 0.0, -10.1, -10.1, -5.8,
|
|
10.1, 10.1, 5.8, -0.5, 0.5, 0.0
|
|
};
|
|
__ lwc1(f2, MemOperand(a0, offsetof(TestFloat, a)) );
|
|
__ lwc1(f4, MemOperand(a0, offsetof(TestFloat, b)) );
|
|
__ Ldc1(f8, MemOperand(a0, offsetof(TestFloat, c)));
|
|
__ Ldc1(f10, MemOperand(a0, offsetof(TestFloat, d)));
|
|
__ sub_s(f6, f2, f4);
|
|
__ sub_d(f12, f8, f10);
|
|
__ swc1(f6, MemOperand(a0, offsetof(TestFloat, resultS)) );
|
|
__ Sdc1(f12, MemOperand(a0, offsetof(TestFloat, resultD)));
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
for (int i = 0; i < kTableLength; i++) {
|
|
test.a = inputfs_S[i];
|
|
test.b = inputft_S[i];
|
|
test.c = inputfs_D[i];
|
|
test.d = inputft_D[i];
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.resultS, outputs_S[i]);
|
|
CHECK_EQ(test.resultD, outputs_D[i]);
|
|
}
|
|
}
|
|
|
|
|
|
TEST(sqrt_rsqrt_recip) {
|
|
const int kTableLength = 4;
|
|
const double deltaDouble = 2E-15;
|
|
const float deltaFloat = 2E-7;
|
|
const float sqrt2_s = sqrt(2);
|
|
const double sqrt2_d = sqrt(2);
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
struct TestFloat {
|
|
float a;
|
|
float resultS;
|
|
float resultS1;
|
|
float resultS2;
|
|
double c;
|
|
double resultD;
|
|
double resultD1;
|
|
double resultD2;
|
|
};
|
|
TestFloat test;
|
|
|
|
double inputs_D[kTableLength] = {
|
|
0.0L, 4.0L, 2.0L, 4e-28L
|
|
};
|
|
|
|
double outputs_D[kTableLength] = {
|
|
0.0L, 2.0L, sqrt2_d, 2e-14L
|
|
};
|
|
float inputs_S[kTableLength] = {
|
|
0.0, 4.0, 2.0, 4e-28
|
|
};
|
|
|
|
float outputs_S[kTableLength] = {
|
|
0.0, 2.0, sqrt2_s, 2e-14
|
|
};
|
|
|
|
|
|
__ lwc1(f2, MemOperand(a0, offsetof(TestFloat, a)) );
|
|
__ Ldc1(f8, MemOperand(a0, offsetof(TestFloat, c)));
|
|
__ sqrt_s(f6, f2);
|
|
__ sqrt_d(f12, f8);
|
|
|
|
if (IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) {
|
|
__ rsqrt_d(f14, f8);
|
|
__ rsqrt_s(f16, f2);
|
|
__ recip_d(f18, f8);
|
|
__ recip_s(f4, f2);
|
|
}
|
|
__ swc1(f6, MemOperand(a0, offsetof(TestFloat, resultS)) );
|
|
__ Sdc1(f12, MemOperand(a0, offsetof(TestFloat, resultD)));
|
|
|
|
if (IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) {
|
|
__ swc1(f16, MemOperand(a0, offsetof(TestFloat, resultS1)) );
|
|
__ Sdc1(f14, MemOperand(a0, offsetof(TestFloat, resultD1)));
|
|
__ swc1(f4, MemOperand(a0, offsetof(TestFloat, resultS2)) );
|
|
__ Sdc1(f18, MemOperand(a0, offsetof(TestFloat, resultD2)));
|
|
}
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
|
|
for (int i = 0; i < kTableLength; i++) {
|
|
float f1;
|
|
double d1;
|
|
test.a = inputs_S[i];
|
|
test.c = inputs_D[i];
|
|
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
|
|
CHECK_EQ(test.resultS, outputs_S[i]);
|
|
CHECK_EQ(test.resultD, outputs_D[i]);
|
|
|
|
if (IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) {
|
|
if (i != 0) {
|
|
f1 = test.resultS1 - 1.0F/outputs_S[i];
|
|
f1 = (f1 < 0) ? f1 : -f1;
|
|
CHECK(f1 <= deltaFloat);
|
|
d1 = test.resultD1 - 1.0L/outputs_D[i];
|
|
d1 = (d1 < 0) ? d1 : -d1;
|
|
CHECK(d1 <= deltaDouble);
|
|
f1 = test.resultS2 - 1.0F/inputs_S[i];
|
|
f1 = (f1 < 0) ? f1 : -f1;
|
|
CHECK(f1 <= deltaFloat);
|
|
d1 = test.resultD2 - 1.0L/inputs_D[i];
|
|
d1 = (d1 < 0) ? d1 : -d1;
|
|
CHECK(d1 <= deltaDouble);
|
|
} else {
|
|
CHECK_EQ(test.resultS1, 1.0F/outputs_S[i]);
|
|
CHECK_EQ(test.resultD1, 1.0L/outputs_D[i]);
|
|
CHECK_EQ(test.resultS2, 1.0F/inputs_S[i]);
|
|
CHECK_EQ(test.resultD2, 1.0L/inputs_D[i]);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
TEST(neg) {
|
|
const int kTableLength = 3;
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
struct TestFloat {
|
|
float a;
|
|
float resultS;
|
|
double c;
|
|
double resultD;
|
|
};
|
|
|
|
TestFloat test;
|
|
double inputs_D[kTableLength] = {
|
|
0.0, 4.0, -2.0
|
|
};
|
|
|
|
double outputs_D[kTableLength] = {
|
|
0.0, -4.0, 2.0
|
|
};
|
|
float inputs_S[kTableLength] = {
|
|
0.0, 4.0, -2.0
|
|
};
|
|
|
|
float outputs_S[kTableLength] = {
|
|
0.0, -4.0, 2.0
|
|
};
|
|
__ lwc1(f2, MemOperand(a0, offsetof(TestFloat, a)) );
|
|
__ Ldc1(f8, MemOperand(a0, offsetof(TestFloat, c)));
|
|
__ neg_s(f6, f2);
|
|
__ neg_d(f12, f8);
|
|
__ swc1(f6, MemOperand(a0, offsetof(TestFloat, resultS)) );
|
|
__ Sdc1(f12, MemOperand(a0, offsetof(TestFloat, resultD)));
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
for (int i = 0; i < kTableLength; i++) {
|
|
test.a = inputs_S[i];
|
|
test.c = inputs_D[i];
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.resultS, outputs_S[i]);
|
|
CHECK_EQ(test.resultD, outputs_D[i]);
|
|
}
|
|
}
|
|
|
|
|
|
TEST(mul) {
|
|
const int kTableLength = 4;
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
struct TestFloat {
|
|
float a;
|
|
float b;
|
|
float resultS;
|
|
double c;
|
|
double d;
|
|
double resultD;
|
|
};
|
|
|
|
TestFloat test;
|
|
double inputfs_D[kTableLength] = {
|
|
5.3, -5.3, 5.3, -2.9
|
|
};
|
|
double inputft_D[kTableLength] = {
|
|
4.8, 4.8, -4.8, -0.29
|
|
};
|
|
|
|
float inputfs_S[kTableLength] = {
|
|
5.3, -5.3, 5.3, -2.9
|
|
};
|
|
float inputft_S[kTableLength] = {
|
|
4.8, 4.8, -4.8, -0.29
|
|
};
|
|
|
|
__ lwc1(f2, MemOperand(a0, offsetof(TestFloat, a)) );
|
|
__ lwc1(f4, MemOperand(a0, offsetof(TestFloat, b)) );
|
|
__ Ldc1(f6, MemOperand(a0, offsetof(TestFloat, c)));
|
|
__ Ldc1(f8, MemOperand(a0, offsetof(TestFloat, d)));
|
|
__ mul_s(f10, f2, f4);
|
|
__ mul_d(f12, f6, f8);
|
|
__ swc1(f10, MemOperand(a0, offsetof(TestFloat, resultS)) );
|
|
__ Sdc1(f12, MemOperand(a0, offsetof(TestFloat, resultD)));
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
for (int i = 0; i < kTableLength; i++) {
|
|
test.a = inputfs_S[i];
|
|
test.b = inputft_S[i];
|
|
test.c = inputfs_D[i];
|
|
test.d = inputft_D[i];
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.resultS, inputfs_S[i]*inputft_S[i]);
|
|
CHECK_EQ(test.resultD, inputfs_D[i]*inputft_D[i]);
|
|
}
|
|
}
|
|
|
|
|
|
TEST(mov) {
|
|
const int kTableLength = 4;
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
struct TestFloat {
|
|
double a;
|
|
double b;
|
|
float c;
|
|
float d;
|
|
};
|
|
|
|
TestFloat test;
|
|
double inputs_D[kTableLength] = {
|
|
5.3, -5.3, 5.3, -2.9
|
|
};
|
|
double inputs_S[kTableLength] = {
|
|
4.8, 4.8, -4.8, -0.29
|
|
};
|
|
|
|
float outputs_S[kTableLength] = {
|
|
4.8, 4.8, -4.8, -0.29
|
|
};
|
|
double outputs_D[kTableLength] = {
|
|
5.3, -5.3, 5.3, -2.9
|
|
};
|
|
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(TestFloat, a)));
|
|
__ lwc1(f6, MemOperand(a0, offsetof(TestFloat, c)) );
|
|
__ mov_s(f8, f6);
|
|
__ mov_d(f10, f4);
|
|
__ swc1(f8, MemOperand(a0, offsetof(TestFloat, d)) );
|
|
__ Sdc1(f10, MemOperand(a0, offsetof(TestFloat, b)));
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
for (int i = 0; i < kTableLength; i++) {
|
|
test.a = inputs_D[i];
|
|
test.c = inputs_S[i];
|
|
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.b, outputs_D[i]);
|
|
CHECK_EQ(test.d, outputs_S[i]);
|
|
}
|
|
}
|
|
|
|
|
|
TEST(floor_w) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
struct Test {
|
|
uint32_t isNaN2008;
|
|
double a;
|
|
float b;
|
|
int32_t c; // a floor result
|
|
int32_t d; // b floor result
|
|
};
|
|
const int kTableLength = 15;
|
|
double inputs_D[kTableLength] = {
|
|
2.1, 2.6, 2.5, 3.1, 3.6, 3.5,
|
|
-2.1, -2.6, -2.5, -3.1, -3.6, -3.5,
|
|
2147483648.0,
|
|
std::numeric_limits<double>::quiet_NaN(),
|
|
std::numeric_limits<double>::infinity()
|
|
};
|
|
float inputs_S[kTableLength] = {
|
|
2.1, 2.6, 2.5, 3.1, 3.6, 3.5,
|
|
-2.1, -2.6, -2.5, -3.1, -3.6, -3.5,
|
|
2147483648.0,
|
|
std::numeric_limits<float>::quiet_NaN(),
|
|
std::numeric_limits<float>::infinity()
|
|
};
|
|
double outputs[kTableLength] = {
|
|
2.0, 2.0, 2.0, 3.0, 3.0, 3.0,
|
|
-3.0, -3.0, -3.0, -4.0, -4.0, -4.0,
|
|
kFPUInvalidResult, kFPUInvalidResult,
|
|
kFPUInvalidResult};
|
|
double outputsNaN2008[kTableLength] = {
|
|
2.0, 2.0, 2.0, 3.0, 3.0, 3.0,
|
|
-3.0, -3.0, -3.0, -4.0, -4.0, -4.0,
|
|
kFPUInvalidResult,
|
|
0,
|
|
kFPUInvalidResult};
|
|
|
|
__ cfc1(t1, FCSR);
|
|
__ sw(t1, MemOperand(a0, offsetof(Test, isNaN2008)));
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(Test, a)));
|
|
__ lwc1(f6, MemOperand(a0, offsetof(Test, b)) );
|
|
__ floor_w_d(f8, f4);
|
|
__ floor_w_s(f10, f6);
|
|
__ swc1(f8, MemOperand(a0, offsetof(Test, c)) );
|
|
__ swc1(f10, MemOperand(a0, offsetof(Test, d)) );
|
|
__ jr(ra);
|
|
__ nop();
|
|
Test test;
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
for (int i = 0; i < kTableLength; i++) {
|
|
test.a = inputs_D[i];
|
|
test.b = inputs_S[i];
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
if ((test.isNaN2008 & kFCSRNaN2008FlagMask) && kArchVariant == kMips32r6) {
|
|
CHECK_EQ(test.c, outputsNaN2008[i]);
|
|
} else {
|
|
CHECK_EQ(test.c, outputs[i]);
|
|
}
|
|
CHECK_EQ(test.d, test.c);
|
|
}
|
|
}
|
|
|
|
|
|
TEST(floor_l) {
|
|
if (IsFp64Mode()) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
const double dFPU64InvalidResult = static_cast<double>(kFPU64InvalidResult);
|
|
struct Test {
|
|
uint32_t isNaN2008;
|
|
double a;
|
|
float b;
|
|
int64_t c;
|
|
int64_t d;
|
|
};
|
|
const int kTableLength = 15;
|
|
double inputs_D[kTableLength] = {
|
|
2.1, 2.6, 2.5, 3.1, 3.6, 3.5,
|
|
-2.1, -2.6, -2.5, -3.1, -3.6, -3.5,
|
|
2147483648.0,
|
|
std::numeric_limits<double>::quiet_NaN(),
|
|
std::numeric_limits<double>::infinity()
|
|
};
|
|
float inputs_S[kTableLength] = {
|
|
2.1, 2.6, 2.5, 3.1, 3.6, 3.5,
|
|
-2.1, -2.6, -2.5, -3.1, -3.6, -3.5,
|
|
2147483648.0,
|
|
std::numeric_limits<float>::quiet_NaN(),
|
|
std::numeric_limits<float>::infinity()
|
|
};
|
|
double outputs[kTableLength] = {
|
|
2.0, 2.0, 2.0, 3.0, 3.0, 3.0,
|
|
-3.0, -3.0, -3.0, -4.0, -4.0, -4.0,
|
|
2147483648.0, dFPU64InvalidResult,
|
|
dFPU64InvalidResult};
|
|
double outputsNaN2008[kTableLength] = {
|
|
2.0, 2.0, 2.0, 3.0, 3.0, 3.0,
|
|
-3.0, -3.0, -3.0, -4.0, -4.0, -4.0,
|
|
2147483648.0,
|
|
0,
|
|
dFPU64InvalidResult};
|
|
|
|
__ cfc1(t1, FCSR);
|
|
__ sw(t1, MemOperand(a0, offsetof(Test, isNaN2008)));
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(Test, a)));
|
|
__ lwc1(f6, MemOperand(a0, offsetof(Test, b)) );
|
|
__ floor_l_d(f8, f4);
|
|
__ floor_l_s(f10, f6);
|
|
__ Sdc1(f8, MemOperand(a0, offsetof(Test, c)));
|
|
__ Sdc1(f10, MemOperand(a0, offsetof(Test, d)));
|
|
__ jr(ra);
|
|
__ nop();
|
|
Test test;
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
for (int i = 0; i < kTableLength; i++) {
|
|
test.a = inputs_D[i];
|
|
test.b = inputs_S[i];
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
if ((test.isNaN2008 & kFCSRNaN2008FlagMask) &&
|
|
kArchVariant == kMips32r6) {
|
|
CHECK_EQ(test.c, outputsNaN2008[i]);
|
|
} else {
|
|
CHECK_EQ(test.c, outputs[i]);
|
|
}
|
|
CHECK_EQ(test.d, test.c);
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
TEST(ceil_w) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
struct Test {
|
|
uint32_t isNaN2008;
|
|
double a;
|
|
float b;
|
|
int32_t c; // a floor result
|
|
int32_t d; // b floor result
|
|
};
|
|
const int kTableLength = 15;
|
|
double inputs_D[kTableLength] = {
|
|
2.1, 2.6, 2.5, 3.1, 3.6, 3.5,
|
|
-2.1, -2.6, -2.5, -3.1, -3.6, -3.5,
|
|
2147483648.0,
|
|
std::numeric_limits<double>::quiet_NaN(),
|
|
std::numeric_limits<double>::infinity()
|
|
};
|
|
float inputs_S[kTableLength] = {
|
|
2.1, 2.6, 2.5, 3.1, 3.6, 3.5,
|
|
-2.1, -2.6, -2.5, -3.1, -3.6, -3.5,
|
|
2147483648.0,
|
|
std::numeric_limits<float>::quiet_NaN(),
|
|
std::numeric_limits<float>::infinity()
|
|
};
|
|
double outputs[kTableLength] = {
|
|
3.0, 3.0, 3.0, 4.0, 4.0, 4.0,
|
|
-2.0, -2.0, -2.0, -3.0, -3.0, -3.0,
|
|
kFPUInvalidResult, kFPUInvalidResult,
|
|
kFPUInvalidResult};
|
|
double outputsNaN2008[kTableLength] = {
|
|
3.0, 3.0, 3.0, 4.0, 4.0, 4.0,
|
|
-2.0, -2.0, -2.0, -3.0, -3.0, -3.0,
|
|
kFPUInvalidResult,
|
|
0,
|
|
kFPUInvalidResult};
|
|
|
|
__ cfc1(t1, FCSR);
|
|
__ sw(t1, MemOperand(a0, offsetof(Test, isNaN2008)));
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(Test, a)));
|
|
__ lwc1(f6, MemOperand(a0, offsetof(Test, b)) );
|
|
__ ceil_w_d(f8, f4);
|
|
__ ceil_w_s(f10, f6);
|
|
__ swc1(f8, MemOperand(a0, offsetof(Test, c)) );
|
|
__ swc1(f10, MemOperand(a0, offsetof(Test, d)) );
|
|
__ jr(ra);
|
|
__ nop();
|
|
Test test;
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
for (int i = 0; i < kTableLength; i++) {
|
|
test.a = inputs_D[i];
|
|
test.b = inputs_S[i];
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
if ((test.isNaN2008 & kFCSRNaN2008FlagMask) && kArchVariant == kMips32r6) {
|
|
CHECK_EQ(test.c, outputsNaN2008[i]);
|
|
} else {
|
|
CHECK_EQ(test.c, outputs[i]);
|
|
}
|
|
CHECK_EQ(test.d, test.c);
|
|
}
|
|
}
|
|
|
|
|
|
TEST(ceil_l) {
|
|
if (IsFp64Mode()) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
const double dFPU64InvalidResult = static_cast<double>(kFPU64InvalidResult);
|
|
struct Test {
|
|
uint32_t isNaN2008;
|
|
double a;
|
|
float b;
|
|
int64_t c;
|
|
int64_t d;
|
|
};
|
|
const int kTableLength = 15;
|
|
double inputs_D[kTableLength] = {
|
|
2.1, 2.6, 2.5, 3.1, 3.6, 3.5,
|
|
-2.1, -2.6, -2.5, -3.1, -3.6, -3.5,
|
|
2147483648.0,
|
|
std::numeric_limits<double>::quiet_NaN(),
|
|
std::numeric_limits<double>::infinity()
|
|
};
|
|
float inputs_S[kTableLength] = {
|
|
2.1, 2.6, 2.5, 3.1, 3.6, 3.5,
|
|
-2.1, -2.6, -2.5, -3.1, -3.6, -3.5,
|
|
2147483648.0,
|
|
std::numeric_limits<float>::quiet_NaN(),
|
|
std::numeric_limits<float>::infinity()
|
|
};
|
|
double outputs[kTableLength] = {
|
|
3.0, 3.0, 3.0, 4.0, 4.0, 4.0,
|
|
-2.0, -2.0, -2.0, -3.0, -3.0, -3.0,
|
|
2147483648.0, dFPU64InvalidResult,
|
|
dFPU64InvalidResult};
|
|
double outputsNaN2008[kTableLength] = {
|
|
3.0, 3.0, 3.0, 4.0, 4.0, 4.0,
|
|
-2.0, -2.0, -2.0, -3.0, -3.0, -3.0,
|
|
2147483648.0,
|
|
0,
|
|
dFPU64InvalidResult};
|
|
|
|
__ cfc1(t1, FCSR);
|
|
__ sw(t1, MemOperand(a0, offsetof(Test, isNaN2008)));
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(Test, a)));
|
|
__ lwc1(f6, MemOperand(a0, offsetof(Test, b)) );
|
|
__ ceil_l_d(f8, f4);
|
|
__ ceil_l_s(f10, f6);
|
|
__ Sdc1(f8, MemOperand(a0, offsetof(Test, c)));
|
|
__ Sdc1(f10, MemOperand(a0, offsetof(Test, d)));
|
|
__ jr(ra);
|
|
__ nop();
|
|
Test test;
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
for (int i = 0; i < kTableLength; i++) {
|
|
test.a = inputs_D[i];
|
|
test.b = inputs_S[i];
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
if ((test.isNaN2008 & kFCSRNaN2008FlagMask) &&
|
|
kArchVariant == kMips32r6) {
|
|
CHECK_EQ(test.c, outputsNaN2008[i]);
|
|
} else {
|
|
CHECK_EQ(test.c, outputs[i]);
|
|
}
|
|
CHECK_EQ(test.d, test.c);
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
TEST(jump_tables1) {
|
|
// Test jump tables with forward jumps.
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
Assembler assm(AssemblerOptions{});
|
|
|
|
const int kNumCases = 512;
|
|
int values[kNumCases];
|
|
isolate->random_number_generator()->NextBytes(values, sizeof(values));
|
|
Label labels[kNumCases];
|
|
|
|
__ addiu(sp, sp, -4);
|
|
__ sw(ra, MemOperand(sp));
|
|
|
|
Label done;
|
|
{
|
|
__ BlockTrampolinePoolFor(kNumCases + 7);
|
|
|
|
__ nal();
|
|
__ nop();
|
|
__ sll(at, a0, 2);
|
|
__ addu(at, at, ra);
|
|
__ lw(at, MemOperand(at, 5 * kInstrSize));
|
|
__ jr(at);
|
|
__ nop();
|
|
for (int i = 0; i < kNumCases; ++i) {
|
|
__ dd(&labels[i]);
|
|
}
|
|
}
|
|
|
|
for (int i = 0; i < kNumCases; ++i) {
|
|
__ bind(&labels[i]);
|
|
__ lui(v0, (values[i] >> 16) & 0xFFFF);
|
|
__ ori(v0, v0, values[i] & 0xFFFF);
|
|
__ b(&done);
|
|
__ nop();
|
|
}
|
|
|
|
__ bind(&done);
|
|
__ lw(ra, MemOperand(sp));
|
|
__ addiu(sp, sp, 4);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CHECK_EQ(0, assm.UnboundLabelsCount());
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
#ifdef OBJECT_PRINT
|
|
code->Print(std::cout);
|
|
#endif
|
|
auto f = GeneratedCode<F1>::FromCode(*code);
|
|
for (int i = 0; i < kNumCases; ++i) {
|
|
int res = reinterpret_cast<int>(f.Call(i, 0, 0, 0, 0));
|
|
::printf("f(%d) = %d\n", i, res);
|
|
CHECK_EQ(values[i], res);
|
|
}
|
|
}
|
|
|
|
|
|
TEST(jump_tables2) {
|
|
// Test jump tables with backward jumps.
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
Assembler assm(AssemblerOptions{});
|
|
|
|
const int kNumCases = 512;
|
|
int values[kNumCases];
|
|
isolate->random_number_generator()->NextBytes(values, sizeof(values));
|
|
Label labels[kNumCases];
|
|
|
|
__ addiu(sp, sp, -4);
|
|
__ sw(ra, MemOperand(sp));
|
|
|
|
Label done, dispatch;
|
|
__ b(&dispatch);
|
|
__ nop();
|
|
|
|
for (int i = 0; i < kNumCases; ++i) {
|
|
__ bind(&labels[i]);
|
|
__ lui(v0, (values[i] >> 16) & 0xFFFF);
|
|
__ ori(v0, v0, values[i] & 0xFFFF);
|
|
__ b(&done);
|
|
__ nop();
|
|
}
|
|
|
|
__ bind(&dispatch);
|
|
{
|
|
__ BlockTrampolinePoolFor(kNumCases + 7);
|
|
|
|
__ nal();
|
|
__ nop();
|
|
__ sll(at, a0, 2);
|
|
__ addu(at, at, ra);
|
|
__ lw(at, MemOperand(at, 5 * kInstrSize));
|
|
__ jr(at);
|
|
__ nop();
|
|
for (int i = 0; i < kNumCases; ++i) {
|
|
__ dd(&labels[i]);
|
|
}
|
|
}
|
|
|
|
__ bind(&done);
|
|
__ lw(ra, MemOperand(sp));
|
|
__ addiu(sp, sp, 4);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
#ifdef OBJECT_PRINT
|
|
code->Print(std::cout);
|
|
#endif
|
|
auto f = GeneratedCode<F1>::FromCode(*code);
|
|
for (int i = 0; i < kNumCases; ++i) {
|
|
int res = reinterpret_cast<int>(f.Call(i, 0, 0, 0, 0));
|
|
::printf("f(%d) = %d\n", i, res);
|
|
CHECK_EQ(values[i], res);
|
|
}
|
|
}
|
|
|
|
|
|
TEST(jump_tables3) {
|
|
// Test jump tables with backward jumps and embedded heap objects.
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
Assembler assm(AssemblerOptions{});
|
|
|
|
const int kNumCases = 256;
|
|
Handle<Object> values[kNumCases];
|
|
for (int i = 0; i < kNumCases; ++i) {
|
|
double value = isolate->random_number_generator()->NextDouble();
|
|
values[i] = isolate->factory()->NewHeapNumber<AllocationType::kOld>(value);
|
|
}
|
|
Label labels[kNumCases];
|
|
Object obj;
|
|
int32_t imm32;
|
|
|
|
__ addiu(sp, sp, -4);
|
|
__ sw(ra, MemOperand(sp));
|
|
|
|
Label done, dispatch;
|
|
__ b(&dispatch);
|
|
|
|
|
|
for (int i = 0; i < kNumCases; ++i) {
|
|
__ bind(&labels[i]);
|
|
obj = *values[i];
|
|
imm32 = obj.ptr();
|
|
__ lui(v0, (imm32 >> 16) & 0xFFFF);
|
|
__ ori(v0, v0, imm32 & 0xFFFF);
|
|
__ b(&done);
|
|
__ nop();
|
|
}
|
|
|
|
__ bind(&dispatch);
|
|
{
|
|
__ BlockTrampolinePoolFor(kNumCases + 7);
|
|
|
|
__ nal();
|
|
__ nop();
|
|
__ sll(at, a0, 2);
|
|
__ addu(at, at, ra);
|
|
__ lw(at, MemOperand(at, 5 * kInstrSize));
|
|
__ jr(at);
|
|
__ nop();
|
|
for (int i = 0; i < kNumCases; ++i) {
|
|
__ dd(&labels[i]);
|
|
}
|
|
}
|
|
|
|
__ bind(&done);
|
|
__ lw(ra, MemOperand(sp));
|
|
__ addiu(sp, sp, 4);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
#ifdef OBJECT_PRINT
|
|
code->Print(std::cout);
|
|
#endif
|
|
auto f = GeneratedCode<F1>::FromCode(*code);
|
|
for (int i = 0; i < kNumCases; ++i) {
|
|
Handle<Object> result(
|
|
Object(reinterpret_cast<Address>(f.Call(i, 0, 0, 0, 0))), isolate);
|
|
#ifdef OBJECT_PRINT
|
|
::printf("f(%d) = ", i);
|
|
result->Print(std::cout);
|
|
::printf("\n");
|
|
#endif
|
|
CHECK(values[i].is_identical_to(result));
|
|
}
|
|
}
|
|
|
|
|
|
TEST(BITSWAP) {
|
|
// Test BITSWAP
|
|
if (IsMipsArchVariant(kMips32r6)) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
struct T {
|
|
int32_t r1;
|
|
int32_t r2;
|
|
int32_t r3;
|
|
int32_t r4;
|
|
};
|
|
T t;
|
|
|
|
Assembler assm(AssemblerOptions{});
|
|
|
|
__ lw(a2, MemOperand(a0, offsetof(T, r1)));
|
|
__ nop();
|
|
__ bitswap(a1, a2);
|
|
__ sw(a1, MemOperand(a0, offsetof(T, r1)));
|
|
|
|
__ lw(a2, MemOperand(a0, offsetof(T, r2)));
|
|
__ nop();
|
|
__ bitswap(a1, a2);
|
|
__ sw(a1, MemOperand(a0, offsetof(T, r2)));
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
t.r1 = 0x781A15C3;
|
|
t.r2 = 0x8B71FCDE;
|
|
f.Call(&t, 0, 0, 0, 0);
|
|
|
|
CHECK_EQ(static_cast<int32_t>(0x1E58A8C3), t.r1);
|
|
CHECK_EQ(static_cast<int32_t>(0xD18E3F7B), t.r2);
|
|
}
|
|
}
|
|
|
|
|
|
TEST(class_fmt) {
|
|
if (IsMipsArchVariant(kMips32r6)) {
|
|
// Test CLASS.fmt instruction.
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
struct T {
|
|
double dSignalingNan;
|
|
double dQuietNan;
|
|
double dNegInf;
|
|
double dNegNorm;
|
|
double dNegSubnorm;
|
|
double dNegZero;
|
|
double dPosInf;
|
|
double dPosNorm;
|
|
double dPosSubnorm;
|
|
double dPosZero;
|
|
float fSignalingNan;
|
|
float fQuietNan;
|
|
float fNegInf;
|
|
float fNegNorm;
|
|
float fNegSubnorm;
|
|
float fNegZero;
|
|
float fPosInf;
|
|
float fPosNorm;
|
|
float fPosSubnorm;
|
|
float fPosZero;
|
|
};
|
|
T t;
|
|
|
|
// Create a function that accepts &t, and loads, manipulates, and stores
|
|
// the doubles t.a ... t.f.
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(T, dSignalingNan)));
|
|
__ class_d(f6, f4);
|
|
__ Sdc1(f6, MemOperand(a0, offsetof(T, dSignalingNan)));
|
|
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(T, dQuietNan)));
|
|
__ class_d(f6, f4);
|
|
__ Sdc1(f6, MemOperand(a0, offsetof(T, dQuietNan)));
|
|
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(T, dNegInf)));
|
|
__ class_d(f6, f4);
|
|
__ Sdc1(f6, MemOperand(a0, offsetof(T, dNegInf)));
|
|
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(T, dNegNorm)));
|
|
__ class_d(f6, f4);
|
|
__ Sdc1(f6, MemOperand(a0, offsetof(T, dNegNorm)));
|
|
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(T, dNegSubnorm)));
|
|
__ class_d(f6, f4);
|
|
__ Sdc1(f6, MemOperand(a0, offsetof(T, dNegSubnorm)));
|
|
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(T, dNegZero)));
|
|
__ class_d(f6, f4);
|
|
__ Sdc1(f6, MemOperand(a0, offsetof(T, dNegZero)));
|
|
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(T, dPosInf)));
|
|
__ class_d(f6, f4);
|
|
__ Sdc1(f6, MemOperand(a0, offsetof(T, dPosInf)));
|
|
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(T, dPosNorm)));
|
|
__ class_d(f6, f4);
|
|
__ Sdc1(f6, MemOperand(a0, offsetof(T, dPosNorm)));
|
|
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(T, dPosSubnorm)));
|
|
__ class_d(f6, f4);
|
|
__ Sdc1(f6, MemOperand(a0, offsetof(T, dPosSubnorm)));
|
|
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(T, dPosZero)));
|
|
__ class_d(f6, f4);
|
|
__ Sdc1(f6, MemOperand(a0, offsetof(T, dPosZero)));
|
|
|
|
// Testing instruction CLASS.S
|
|
__ lwc1(f4, MemOperand(a0, offsetof(T, fSignalingNan)));
|
|
__ class_s(f6, f4);
|
|
__ swc1(f6, MemOperand(a0, offsetof(T, fSignalingNan)));
|
|
|
|
__ lwc1(f4, MemOperand(a0, offsetof(T, fQuietNan)));
|
|
__ class_s(f6, f4);
|
|
__ swc1(f6, MemOperand(a0, offsetof(T, fQuietNan)));
|
|
|
|
__ lwc1(f4, MemOperand(a0, offsetof(T, fNegInf)));
|
|
__ class_s(f6, f4);
|
|
__ swc1(f6, MemOperand(a0, offsetof(T, fNegInf)));
|
|
|
|
__ lwc1(f4, MemOperand(a0, offsetof(T, fNegNorm)));
|
|
__ class_s(f6, f4);
|
|
__ swc1(f6, MemOperand(a0, offsetof(T, fNegNorm)));
|
|
|
|
__ lwc1(f4, MemOperand(a0, offsetof(T, fNegSubnorm)));
|
|
__ class_s(f6, f4);
|
|
__ swc1(f6, MemOperand(a0, offsetof(T, fNegSubnorm)));
|
|
|
|
__ lwc1(f4, MemOperand(a0, offsetof(T, fNegZero)));
|
|
__ class_s(f6, f4);
|
|
__ swc1(f6, MemOperand(a0, offsetof(T, fNegZero)));
|
|
|
|
__ lwc1(f4, MemOperand(a0, offsetof(T, fPosInf)));
|
|
__ class_s(f6, f4);
|
|
__ swc1(f6, MemOperand(a0, offsetof(T, fPosInf)));
|
|
|
|
__ lwc1(f4, MemOperand(a0, offsetof(T, fPosNorm)));
|
|
__ class_s(f6, f4);
|
|
__ swc1(f6, MemOperand(a0, offsetof(T, fPosNorm)));
|
|
|
|
__ lwc1(f4, MemOperand(a0, offsetof(T, fPosSubnorm)));
|
|
__ class_s(f6, f4);
|
|
__ swc1(f6, MemOperand(a0, offsetof(T, fPosSubnorm)));
|
|
|
|
__ lwc1(f4, MemOperand(a0, offsetof(T, fPosZero)));
|
|
__ class_s(f6, f4);
|
|
__ swc1(f6, MemOperand(a0, offsetof(T, fPosZero)));
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
|
|
t.dSignalingNan = std::numeric_limits<double>::signaling_NaN();
|
|
t.dQuietNan = std::numeric_limits<double>::quiet_NaN();
|
|
t.dNegInf = -1.0 / 0.0;
|
|
t.dNegNorm = -5.0;
|
|
t.dNegSubnorm = -DBL_MIN / 2.0;
|
|
t.dNegZero = -0.0;
|
|
t.dPosInf = 2.0 / 0.0;
|
|
t.dPosNorm = 275.35;
|
|
t.dPosSubnorm = DBL_MIN / 2.0;
|
|
t.dPosZero = +0.0;
|
|
// Float test values
|
|
|
|
t.fSignalingNan = std::numeric_limits<float>::signaling_NaN();
|
|
t.fQuietNan = std::numeric_limits<float>::quiet_NaN();
|
|
t.fNegInf = -0.5/0.0;
|
|
t.fNegNorm = -FLT_MIN;
|
|
t.fNegSubnorm = -FLT_MIN / 1.5;
|
|
t.fNegZero = -0.0;
|
|
t.fPosInf = 100000.0 / 0.0;
|
|
t.fPosNorm = FLT_MAX;
|
|
t.fPosSubnorm = FLT_MIN / 20.0;
|
|
t.fPosZero = +0.0;
|
|
|
|
f.Call(&t, 0, 0, 0, 0);
|
|
// Expected double results.
|
|
CHECK_EQ(bit_cast<int64_t>(t.dSignalingNan), 0x001);
|
|
CHECK_EQ(bit_cast<int64_t>(t.dQuietNan), 0x002);
|
|
CHECK_EQ(bit_cast<int64_t>(t.dNegInf), 0x004);
|
|
CHECK_EQ(bit_cast<int64_t>(t.dNegNorm), 0x008);
|
|
CHECK_EQ(bit_cast<int64_t>(t.dNegSubnorm), 0x010);
|
|
CHECK_EQ(bit_cast<int64_t>(t.dNegZero), 0x020);
|
|
CHECK_EQ(bit_cast<int64_t>(t.dPosInf), 0x040);
|
|
CHECK_EQ(bit_cast<int64_t>(t.dPosNorm), 0x080);
|
|
CHECK_EQ(bit_cast<int64_t>(t.dPosSubnorm), 0x100);
|
|
CHECK_EQ(bit_cast<int64_t>(t.dPosZero), 0x200);
|
|
|
|
// Expected float results.
|
|
CHECK_EQ(bit_cast<int32_t>(t.fSignalingNan), 0x001);
|
|
CHECK_EQ(bit_cast<int32_t>(t.fQuietNan), 0x002);
|
|
CHECK_EQ(bit_cast<int32_t>(t.fNegInf), 0x004);
|
|
CHECK_EQ(bit_cast<int32_t>(t.fNegNorm), 0x008);
|
|
CHECK_EQ(bit_cast<int32_t>(t.fNegSubnorm), 0x010);
|
|
CHECK_EQ(bit_cast<int32_t>(t.fNegZero), 0x020);
|
|
CHECK_EQ(bit_cast<int32_t>(t.fPosInf), 0x040);
|
|
CHECK_EQ(bit_cast<int32_t>(t.fPosNorm), 0x080);
|
|
CHECK_EQ(bit_cast<int32_t>(t.fPosSubnorm), 0x100);
|
|
CHECK_EQ(bit_cast<int32_t>(t.fPosZero), 0x200);
|
|
}
|
|
}
|
|
|
|
|
|
TEST(ABS) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
struct TestFloat {
|
|
int64_t fir;
|
|
double a;
|
|
float b;
|
|
double fcsr;
|
|
};
|
|
|
|
TestFloat test;
|
|
|
|
// Save FIR.
|
|
__ cfc1(a1, FCSR);
|
|
// Disable FPU exceptions.
|
|
__ ctc1(zero_reg, FCSR);
|
|
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(TestFloat, a)));
|
|
__ abs_d(f10, f4);
|
|
__ Sdc1(f10, MemOperand(a0, offsetof(TestFloat, a)));
|
|
|
|
__ lwc1(f4, MemOperand(a0, offsetof(TestFloat, b)));
|
|
__ abs_s(f10, f4);
|
|
__ swc1(f10, MemOperand(a0, offsetof(TestFloat, b)));
|
|
|
|
// Restore FCSR.
|
|
__ ctc1(a1, FCSR);
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
test.a = -2.0;
|
|
test.b = -2.0;
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.a, 2.0);
|
|
CHECK_EQ(test.b, 2.0);
|
|
|
|
test.a = 2.0;
|
|
test.b = 2.0;
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.a, 2.0);
|
|
CHECK_EQ(test.b, 2.0);
|
|
|
|
// Testing biggest positive number
|
|
test.a = std::numeric_limits<double>::max();
|
|
test.b = std::numeric_limits<float>::max();
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.a, std::numeric_limits<double>::max());
|
|
CHECK_EQ(test.b, std::numeric_limits<float>::max());
|
|
|
|
// Testing smallest negative number
|
|
test.a = -std::numeric_limits<double>::max(); // lowest()
|
|
test.b = -std::numeric_limits<float>::max(); // lowest()
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.a, std::numeric_limits<double>::max());
|
|
CHECK_EQ(test.b, std::numeric_limits<float>::max());
|
|
|
|
// Testing smallest positive number
|
|
test.a = -std::numeric_limits<double>::min();
|
|
test.b = -std::numeric_limits<float>::min();
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.a, std::numeric_limits<double>::min());
|
|
CHECK_EQ(test.b, std::numeric_limits<float>::min());
|
|
|
|
// Testing infinity
|
|
test.a = -std::numeric_limits<double>::max()
|
|
/ std::numeric_limits<double>::min();
|
|
test.b = -std::numeric_limits<float>::max()
|
|
/ std::numeric_limits<float>::min();
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.a, std::numeric_limits<double>::max()
|
|
/ std::numeric_limits<double>::min());
|
|
CHECK_EQ(test.b, std::numeric_limits<float>::max()
|
|
/ std::numeric_limits<float>::min());
|
|
|
|
test.a = std::numeric_limits<double>::quiet_NaN();
|
|
test.b = std::numeric_limits<float>::quiet_NaN();
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK(std::isnan(test.a));
|
|
CHECK(std::isnan(test.b));
|
|
|
|
test.a = std::numeric_limits<double>::signaling_NaN();
|
|
test.b = std::numeric_limits<float>::signaling_NaN();
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK(std::isnan(test.a));
|
|
CHECK(std::isnan(test.b));
|
|
}
|
|
|
|
|
|
TEST(ADD_FMT) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
struct TestFloat {
|
|
double a;
|
|
double b;
|
|
double c;
|
|
float fa;
|
|
float fb;
|
|
float fc;
|
|
};
|
|
|
|
TestFloat test;
|
|
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(TestFloat, a)));
|
|
__ Ldc1(f8, MemOperand(a0, offsetof(TestFloat, b)));
|
|
__ add_d(f10, f8, f4);
|
|
__ Sdc1(f10, MemOperand(a0, offsetof(TestFloat, c)));
|
|
|
|
__ lwc1(f4, MemOperand(a0, offsetof(TestFloat, fa)));
|
|
__ lwc1(f8, MemOperand(a0, offsetof(TestFloat, fb)));
|
|
__ add_s(f10, f8, f4);
|
|
__ swc1(f10, MemOperand(a0, offsetof(TestFloat, fc)));
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
test.a = 2.0;
|
|
test.b = 3.0;
|
|
test.fa = 2.0;
|
|
test.fb = 3.0;
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.c, 5.0);
|
|
CHECK_EQ(test.fc, 5.0);
|
|
|
|
test.a = std::numeric_limits<double>::max();
|
|
test.b = -std::numeric_limits<double>::max(); // lowest()
|
|
test.fa = std::numeric_limits<float>::max();
|
|
test.fb = -std::numeric_limits<float>::max(); // lowest()
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.c, 0.0);
|
|
CHECK_EQ(test.fc, 0.0);
|
|
|
|
test.a = std::numeric_limits<double>::max();
|
|
test.b = std::numeric_limits<double>::max();
|
|
test.fa = std::numeric_limits<float>::max();
|
|
test.fb = std::numeric_limits<float>::max();
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK(!std::isfinite(test.c));
|
|
CHECK(!std::isfinite(test.fc));
|
|
|
|
test.a = 5.0;
|
|
test.b = std::numeric_limits<double>::signaling_NaN();
|
|
test.fa = 5.0;
|
|
test.fb = std::numeric_limits<float>::signaling_NaN();
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK(std::isnan(test.c));
|
|
CHECK(std::isnan(test.fc));
|
|
}
|
|
|
|
|
|
TEST(C_COND_FMT) {
|
|
if ((IsMipsArchVariant(kMips32r1)) || (IsMipsArchVariant(kMips32r2))) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
struct TestFloat {
|
|
double dOp1;
|
|
double dOp2;
|
|
uint32_t dF;
|
|
uint32_t dUn;
|
|
uint32_t dEq;
|
|
uint32_t dUeq;
|
|
uint32_t dOlt;
|
|
uint32_t dUlt;
|
|
uint32_t dOle;
|
|
uint32_t dUle;
|
|
float fOp1;
|
|
float fOp2;
|
|
uint32_t fF;
|
|
uint32_t fUn;
|
|
uint32_t fEq;
|
|
uint32_t fUeq;
|
|
uint32_t fOlt;
|
|
uint32_t fUlt;
|
|
uint32_t fOle;
|
|
uint32_t fUle;
|
|
};
|
|
|
|
TestFloat test;
|
|
|
|
__ li(t1, 1);
|
|
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(TestFloat, dOp1)));
|
|
__ Ldc1(f6, MemOperand(a0, offsetof(TestFloat, dOp2)));
|
|
|
|
__ lwc1(f14, MemOperand(a0, offsetof(TestFloat, fOp1)));
|
|
__ lwc1(f16, MemOperand(a0, offsetof(TestFloat, fOp2)));
|
|
|
|
__ mov(t2, zero_reg);
|
|
__ mov(t3, zero_reg);
|
|
__ c_d(F, f4, f6, 0);
|
|
__ c_s(F, f14, f16, 2);
|
|
__ movt(t2, t1, 0);
|
|
__ movt(t3, t1, 2);
|
|
__ sw(t2, MemOperand(a0, offsetof(TestFloat, dF)) );
|
|
__ sw(t3, MemOperand(a0, offsetof(TestFloat, fF)) );
|
|
|
|
__ mov(t2, zero_reg);
|
|
__ mov(t3, zero_reg);
|
|
__ c_d(UN, f4, f6, 2);
|
|
__ c_s(UN, f14, f16, 4);
|
|
__ movt(t2, t1, 2);
|
|
__ movt(t3, t1, 4);
|
|
__ sw(t2, MemOperand(a0, offsetof(TestFloat, dUn)) );
|
|
__ sw(t3, MemOperand(a0, offsetof(TestFloat, fUn)) );
|
|
|
|
__ mov(t2, zero_reg);
|
|
__ mov(t3, zero_reg);
|
|
__ c_d(EQ, f4, f6, 4);
|
|
__ c_s(EQ, f14, f16, 6);
|
|
__ movt(t2, t1, 4);
|
|
__ movt(t3, t1, 6);
|
|
__ sw(t2, MemOperand(a0, offsetof(TestFloat, dEq)) );
|
|
__ sw(t3, MemOperand(a0, offsetof(TestFloat, fEq)) );
|
|
|
|
__ mov(t2, zero_reg);
|
|
__ mov(t3, zero_reg);
|
|
__ c_d(UEQ, f4, f6, 6);
|
|
__ c_s(UEQ, f14, f16, 0);
|
|
__ movt(t2, t1, 6);
|
|
__ movt(t3, t1, 0);
|
|
__ sw(t2, MemOperand(a0, offsetof(TestFloat, dUeq)) );
|
|
__ sw(t3, MemOperand(a0, offsetof(TestFloat, fUeq)) );
|
|
|
|
__ mov(t2, zero_reg);
|
|
__ mov(t3, zero_reg);
|
|
__ c_d(OLT, f4, f6, 0);
|
|
__ c_s(OLT, f14, f16, 2);
|
|
__ movt(t2, t1, 0);
|
|
__ movt(t3, t1, 2);
|
|
__ sw(t2, MemOperand(a0, offsetof(TestFloat, dOlt)) );
|
|
__ sw(t3, MemOperand(a0, offsetof(TestFloat, fOlt)) );
|
|
|
|
__ mov(t2, zero_reg);
|
|
__ mov(t3, zero_reg);
|
|
__ c_d(ULT, f4, f6, 2);
|
|
__ c_s(ULT, f14, f16, 4);
|
|
__ movt(t2, t1, 2);
|
|
__ movt(t3, t1, 4);
|
|
__ sw(t2, MemOperand(a0, offsetof(TestFloat, dUlt)) );
|
|
__ sw(t3, MemOperand(a0, offsetof(TestFloat, fUlt)) );
|
|
|
|
__ mov(t2, zero_reg);
|
|
__ mov(t3, zero_reg);
|
|
__ c_d(OLE, f4, f6, 4);
|
|
__ c_s(OLE, f14, f16, 6);
|
|
__ movt(t2, t1, 4);
|
|
__ movt(t3, t1, 6);
|
|
__ sw(t2, MemOperand(a0, offsetof(TestFloat, dOle)) );
|
|
__ sw(t3, MemOperand(a0, offsetof(TestFloat, fOle)) );
|
|
|
|
__ mov(t2, zero_reg);
|
|
__ mov(t3, zero_reg);
|
|
__ c_d(ULE, f4, f6, 6);
|
|
__ c_s(ULE, f14, f16, 0);
|
|
__ movt(t2, t1, 6);
|
|
__ movt(t3, t1, 0);
|
|
__ sw(t2, MemOperand(a0, offsetof(TestFloat, dUle)) );
|
|
__ sw(t3, MemOperand(a0, offsetof(TestFloat, fUle)) );
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
test.dOp1 = 2.0;
|
|
test.dOp2 = 3.0;
|
|
test.fOp1 = 2.0;
|
|
test.fOp2 = 3.0;
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.dF, 0U);
|
|
CHECK_EQ(test.dUn, 0U);
|
|
CHECK_EQ(test.dEq, 0U);
|
|
CHECK_EQ(test.dUeq, 0U);
|
|
CHECK_EQ(test.dOlt, 1U);
|
|
CHECK_EQ(test.dUlt, 1U);
|
|
CHECK_EQ(test.dOle, 1U);
|
|
CHECK_EQ(test.dUle, 1U);
|
|
CHECK_EQ(test.fF, 0U);
|
|
CHECK_EQ(test.fUn, 0U);
|
|
CHECK_EQ(test.fEq, 0U);
|
|
CHECK_EQ(test.fUeq, 0U);
|
|
CHECK_EQ(test.fOlt, 1U);
|
|
CHECK_EQ(test.fUlt, 1U);
|
|
CHECK_EQ(test.fOle, 1U);
|
|
CHECK_EQ(test.fUle, 1U);
|
|
|
|
test.dOp1 = std::numeric_limits<double>::max();
|
|
test.dOp2 = std::numeric_limits<double>::min();
|
|
test.fOp1 = std::numeric_limits<float>::min();
|
|
test.fOp2 = -std::numeric_limits<float>::max(); // lowest()
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.dF, 0U);
|
|
CHECK_EQ(test.dUn, 0U);
|
|
CHECK_EQ(test.dEq, 0U);
|
|
CHECK_EQ(test.dUeq, 0U);
|
|
CHECK_EQ(test.dOlt, 0U);
|
|
CHECK_EQ(test.dUlt, 0U);
|
|
CHECK_EQ(test.dOle, 0U);
|
|
CHECK_EQ(test.dUle, 0U);
|
|
CHECK_EQ(test.fF, 0U);
|
|
CHECK_EQ(test.fUn, 0U);
|
|
CHECK_EQ(test.fEq, 0U);
|
|
CHECK_EQ(test.fUeq, 0U);
|
|
CHECK_EQ(test.fOlt, 0U);
|
|
CHECK_EQ(test.fUlt, 0U);
|
|
CHECK_EQ(test.fOle, 0U);
|
|
CHECK_EQ(test.fUle, 0U);
|
|
|
|
test.dOp1 = -std::numeric_limits<double>::max(); // lowest()
|
|
test.dOp2 = -std::numeric_limits<double>::max(); // lowest()
|
|
test.fOp1 = std::numeric_limits<float>::max();
|
|
test.fOp2 = std::numeric_limits<float>::max();
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.dF, 0U);
|
|
CHECK_EQ(test.dUn, 0U);
|
|
CHECK_EQ(test.dEq, 1U);
|
|
CHECK_EQ(test.dUeq, 1U);
|
|
CHECK_EQ(test.dOlt, 0U);
|
|
CHECK_EQ(test.dUlt, 0U);
|
|
CHECK_EQ(test.dOle, 1U);
|
|
CHECK_EQ(test.dUle, 1U);
|
|
CHECK_EQ(test.fF, 0U);
|
|
CHECK_EQ(test.fUn, 0U);
|
|
CHECK_EQ(test.fEq, 1U);
|
|
CHECK_EQ(test.fUeq, 1U);
|
|
CHECK_EQ(test.fOlt, 0U);
|
|
CHECK_EQ(test.fUlt, 0U);
|
|
CHECK_EQ(test.fOle, 1U);
|
|
CHECK_EQ(test.fUle, 1U);
|
|
|
|
test.dOp1 = std::numeric_limits<double>::quiet_NaN();
|
|
test.dOp2 = 0.0;
|
|
test.fOp1 = std::numeric_limits<float>::quiet_NaN();
|
|
test.fOp2 = 0.0;
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.dF, 0U);
|
|
CHECK_EQ(test.dUn, 1U);
|
|
CHECK_EQ(test.dEq, 0U);
|
|
CHECK_EQ(test.dUeq, 1U);
|
|
CHECK_EQ(test.dOlt, 0U);
|
|
CHECK_EQ(test.dUlt, 1U);
|
|
CHECK_EQ(test.dOle, 0U);
|
|
CHECK_EQ(test.dUle, 1U);
|
|
CHECK_EQ(test.fF, 0U);
|
|
CHECK_EQ(test.fUn, 1U);
|
|
CHECK_EQ(test.fEq, 0U);
|
|
CHECK_EQ(test.fUeq, 1U);
|
|
CHECK_EQ(test.fOlt, 0U);
|
|
CHECK_EQ(test.fUlt, 1U);
|
|
CHECK_EQ(test.fOle, 0U);
|
|
CHECK_EQ(test.fUle, 1U);
|
|
}
|
|
}
|
|
|
|
|
|
TEST(CMP_COND_FMT) {
|
|
if (IsMipsArchVariant(kMips32r6)) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
struct TestFloat {
|
|
double dOp1;
|
|
double dOp2;
|
|
double dF;
|
|
double dUn;
|
|
double dEq;
|
|
double dUeq;
|
|
double dOlt;
|
|
double dUlt;
|
|
double dOle;
|
|
double dUle;
|
|
double dOr;
|
|
double dUne;
|
|
double dNe;
|
|
float fOp1;
|
|
float fOp2;
|
|
float fF;
|
|
float fUn;
|
|
float fEq;
|
|
float fUeq;
|
|
float fOlt;
|
|
float fUlt;
|
|
float fOle;
|
|
float fUle;
|
|
float fOr;
|
|
float fUne;
|
|
float fNe;
|
|
};
|
|
|
|
TestFloat test;
|
|
|
|
__ li(t1, 1);
|
|
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(TestFloat, dOp1)));
|
|
__ Ldc1(f6, MemOperand(a0, offsetof(TestFloat, dOp2)));
|
|
|
|
__ lwc1(f14, MemOperand(a0, offsetof(TestFloat, fOp1)));
|
|
__ lwc1(f16, MemOperand(a0, offsetof(TestFloat, fOp2)));
|
|
|
|
__ cmp_d(F, f2, f4, f6);
|
|
__ cmp_s(F, f12, f14, f16);
|
|
__ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dF)));
|
|
__ swc1(f12, MemOperand(a0, offsetof(TestFloat, fF)) );
|
|
|
|
__ cmp_d(UN, f2, f4, f6);
|
|
__ cmp_s(UN, f12, f14, f16);
|
|
__ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dUn)));
|
|
__ swc1(f12, MemOperand(a0, offsetof(TestFloat, fUn)) );
|
|
|
|
__ cmp_d(EQ, f2, f4, f6);
|
|
__ cmp_s(EQ, f12, f14, f16);
|
|
__ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dEq)));
|
|
__ swc1(f12, MemOperand(a0, offsetof(TestFloat, fEq)) );
|
|
|
|
__ cmp_d(UEQ, f2, f4, f6);
|
|
__ cmp_s(UEQ, f12, f14, f16);
|
|
__ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dUeq)));
|
|
__ swc1(f12, MemOperand(a0, offsetof(TestFloat, fUeq)) );
|
|
|
|
__ cmp_d(LT, f2, f4, f6);
|
|
__ cmp_s(LT, f12, f14, f16);
|
|
__ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dOlt)));
|
|
__ swc1(f12, MemOperand(a0, offsetof(TestFloat, fOlt)) );
|
|
|
|
__ cmp_d(ULT, f2, f4, f6);
|
|
__ cmp_s(ULT, f12, f14, f16);
|
|
__ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dUlt)));
|
|
__ swc1(f12, MemOperand(a0, offsetof(TestFloat, fUlt)) );
|
|
|
|
__ cmp_d(LE, f2, f4, f6);
|
|
__ cmp_s(LE, f12, f14, f16);
|
|
__ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dOle)));
|
|
__ swc1(f12, MemOperand(a0, offsetof(TestFloat, fOle)) );
|
|
|
|
__ cmp_d(ULE, f2, f4, f6);
|
|
__ cmp_s(ULE, f12, f14, f16);
|
|
__ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dUle)));
|
|
__ swc1(f12, MemOperand(a0, offsetof(TestFloat, fUle)) );
|
|
|
|
__ cmp_d(ORD, f2, f4, f6);
|
|
__ cmp_s(ORD, f12, f14, f16);
|
|
__ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dOr)));
|
|
__ swc1(f12, MemOperand(a0, offsetof(TestFloat, fOr)) );
|
|
|
|
__ cmp_d(UNE, f2, f4, f6);
|
|
__ cmp_s(UNE, f12, f14, f16);
|
|
__ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dUne)));
|
|
__ swc1(f12, MemOperand(a0, offsetof(TestFloat, fUne)) );
|
|
|
|
__ cmp_d(NE, f2, f4, f6);
|
|
__ cmp_s(NE, f12, f14, f16);
|
|
__ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dNe)));
|
|
__ swc1(f12, MemOperand(a0, offsetof(TestFloat, fNe)) );
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
uint64_t dTrue = 0xFFFFFFFFFFFFFFFF;
|
|
uint64_t dFalse = 0x0000000000000000;
|
|
uint32_t fTrue = 0xFFFFFFFF;
|
|
uint32_t fFalse = 0x00000000;
|
|
|
|
test.dOp1 = 2.0;
|
|
test.dOp2 = 3.0;
|
|
test.fOp1 = 2.0;
|
|
test.fOp2 = 3.0;
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dF), dFalse);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dUn), dFalse);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dEq), dFalse);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dUeq), dFalse);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dOlt), dTrue);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dUlt), dTrue);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dOle), dTrue);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dUle), dTrue);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dOr), dTrue);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dUne), dTrue);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dNe), dTrue);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fF), fFalse);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fUn), fFalse);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fEq), fFalse);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fUeq), fFalse);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fOlt), fTrue);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fUlt), fTrue);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fOle), fTrue);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fUle), fTrue);
|
|
|
|
test.dOp1 = std::numeric_limits<double>::max();
|
|
test.dOp2 = std::numeric_limits<double>::min();
|
|
test.fOp1 = std::numeric_limits<float>::min();
|
|
test.fOp2 = -std::numeric_limits<float>::max(); // lowest()
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dF), dFalse);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dUn), dFalse);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dEq), dFalse);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dUeq), dFalse);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dOlt), dFalse);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dUlt), dFalse);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dOle), dFalse);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dUle), dFalse);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dOr), dTrue);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dUne), dTrue);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dNe), dTrue);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fF), fFalse);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fUn), fFalse);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fEq), fFalse);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fUeq), fFalse);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fOlt), fFalse);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fUlt), fFalse);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fOle), fFalse);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fUle), fFalse);
|
|
|
|
test.dOp1 = -std::numeric_limits<double>::max(); // lowest()
|
|
test.dOp2 = -std::numeric_limits<double>::max(); // lowest()
|
|
test.fOp1 = std::numeric_limits<float>::max();
|
|
test.fOp2 = std::numeric_limits<float>::max();
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dF), dFalse);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dUn), dFalse);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dEq), dTrue);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dUeq), dTrue);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dOlt), dFalse);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dUlt), dFalse);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dOle), dTrue);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dUle), dTrue);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dOr), dTrue);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dUne), dFalse);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dNe), dFalse);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fF), fFalse);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fUn), fFalse);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fEq), fTrue);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fUeq), fTrue);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fOlt), fFalse);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fUlt), fFalse);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fOle), fTrue);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fUle), fTrue);
|
|
|
|
test.dOp1 = std::numeric_limits<double>::quiet_NaN();
|
|
test.dOp2 = 0.0;
|
|
test.fOp1 = std::numeric_limits<float>::quiet_NaN();
|
|
test.fOp2 = 0.0;
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dF), dFalse);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dUn), dTrue);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dEq), dFalse);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dUeq), dTrue);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dOlt), dFalse);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dUlt), dTrue);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dOle), dFalse);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dUle), dTrue);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dOr), dFalse);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dUne), dTrue);
|
|
CHECK_EQ(bit_cast<uint64_t>(test.dNe), dFalse);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fF), fFalse);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fUn), fTrue);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fEq), fFalse);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fUeq), fTrue);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fOlt), fFalse);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fUlt), fTrue);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fOle), fFalse);
|
|
CHECK_EQ(bit_cast<uint32_t>(test.fUle), fTrue);
|
|
}
|
|
}
|
|
|
|
|
|
TEST(CVT) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
struct TestFloat {
|
|
float cvt_d_s_in;
|
|
double cvt_d_s_out;
|
|
int32_t cvt_d_w_in;
|
|
double cvt_d_w_out;
|
|
int64_t cvt_d_l_in;
|
|
double cvt_d_l_out;
|
|
|
|
float cvt_l_s_in;
|
|
int64_t cvt_l_s_out;
|
|
double cvt_l_d_in;
|
|
int64_t cvt_l_d_out;
|
|
|
|
double cvt_s_d_in;
|
|
float cvt_s_d_out;
|
|
int32_t cvt_s_w_in;
|
|
float cvt_s_w_out;
|
|
int64_t cvt_s_l_in;
|
|
float cvt_s_l_out;
|
|
|
|
float cvt_w_s_in;
|
|
int32_t cvt_w_s_out;
|
|
double cvt_w_d_in;
|
|
int32_t cvt_w_d_out;
|
|
};
|
|
|
|
TestFloat test;
|
|
|
|
// Save FCSR.
|
|
__ cfc1(a1, FCSR);
|
|
// Disable FPU exceptions.
|
|
__ ctc1(zero_reg, FCSR);
|
|
|
|
#define GENERATE_CVT_TEST(x, y, z) \
|
|
__ y##c1(f0, MemOperand(a0, offsetof(TestFloat, x##_in))); \
|
|
__ x(f0, f0); \
|
|
__ nop(); \
|
|
__ z##c1(f0, MemOperand(a0, offsetof(TestFloat, x##_out)));
|
|
|
|
GENERATE_CVT_TEST(cvt_d_s, lw, Sd)
|
|
GENERATE_CVT_TEST(cvt_d_w, lw, Sd)
|
|
if ((IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) &&
|
|
IsFp64Mode()) {
|
|
GENERATE_CVT_TEST(cvt_d_l, Ld, Sd)
|
|
}
|
|
|
|
if (IsFp64Mode()) {
|
|
GENERATE_CVT_TEST(cvt_l_s, lw, Sd)
|
|
GENERATE_CVT_TEST(cvt_l_d, Ld, Sd)
|
|
}
|
|
|
|
GENERATE_CVT_TEST(cvt_s_d, Ld, sw)
|
|
GENERATE_CVT_TEST(cvt_s_w, lw, sw)
|
|
if ((IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) &&
|
|
IsFp64Mode()) {
|
|
GENERATE_CVT_TEST(cvt_s_l, Ld, sw)
|
|
}
|
|
|
|
GENERATE_CVT_TEST(cvt_w_s, lw, sw)
|
|
GENERATE_CVT_TEST(cvt_w_d, Ld, sw)
|
|
|
|
// Restore FCSR.
|
|
__ ctc1(a1, FCSR);
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
|
|
test.cvt_d_s_in = -0.51;
|
|
test.cvt_d_w_in = -1;
|
|
test.cvt_d_l_in = -1;
|
|
test.cvt_l_s_in = -0.51;
|
|
test.cvt_l_d_in = -0.51;
|
|
test.cvt_s_d_in = -0.51;
|
|
test.cvt_s_w_in = -1;
|
|
test.cvt_s_l_in = -1;
|
|
test.cvt_w_s_in = -0.51;
|
|
test.cvt_w_d_in = -0.51;
|
|
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.cvt_d_s_out, static_cast<double>(test.cvt_d_s_in));
|
|
CHECK_EQ(test.cvt_d_w_out, static_cast<double>(test.cvt_d_w_in));
|
|
if ((IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) &&
|
|
IsFp64Mode()) {
|
|
CHECK_EQ(test.cvt_d_l_out, static_cast<double>(test.cvt_d_l_in));
|
|
}
|
|
if (IsFp64Mode()) {
|
|
CHECK_EQ(-1, test.cvt_l_s_out);
|
|
CHECK_EQ(-1, test.cvt_l_d_out);
|
|
}
|
|
CHECK_EQ(test.cvt_s_d_out, static_cast<float>(test.cvt_s_d_in));
|
|
CHECK_EQ(test.cvt_s_w_out, static_cast<float>(test.cvt_s_w_in));
|
|
if ((IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) &&
|
|
IsFp64Mode()) {
|
|
CHECK_EQ(test.cvt_s_l_out, static_cast<float>(test.cvt_s_l_in));
|
|
}
|
|
CHECK_EQ(-1, test.cvt_w_s_out);
|
|
CHECK_EQ(-1, test.cvt_w_d_out);
|
|
|
|
test.cvt_d_s_in = 0.49;
|
|
test.cvt_d_w_in = 1;
|
|
test.cvt_d_l_in = 1;
|
|
test.cvt_l_s_in = 0.49;
|
|
test.cvt_l_d_in = 0.49;
|
|
test.cvt_s_d_in = 0.49;
|
|
test.cvt_s_w_in = 1;
|
|
test.cvt_s_l_in = 1;
|
|
test.cvt_w_s_in = 0.49;
|
|
test.cvt_w_d_in = 0.49;
|
|
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.cvt_d_s_out, static_cast<double>(test.cvt_d_s_in));
|
|
CHECK_EQ(test.cvt_d_w_out, static_cast<double>(test.cvt_d_w_in));
|
|
if ((IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) &&
|
|
IsFp64Mode()) {
|
|
CHECK_EQ(test.cvt_d_l_out, static_cast<double>(test.cvt_d_l_in));
|
|
}
|
|
if (IsFp64Mode()) {
|
|
CHECK_EQ(0, test.cvt_l_s_out);
|
|
CHECK_EQ(0, test.cvt_l_d_out);
|
|
}
|
|
CHECK_EQ(test.cvt_s_d_out, static_cast<float>(test.cvt_s_d_in));
|
|
CHECK_EQ(test.cvt_s_w_out, static_cast<float>(test.cvt_s_w_in));
|
|
if ((IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) &&
|
|
IsFp64Mode()) {
|
|
CHECK_EQ(test.cvt_s_l_out, static_cast<float>(test.cvt_s_l_in));
|
|
}
|
|
CHECK_EQ(0, test.cvt_w_s_out);
|
|
CHECK_EQ(0, test.cvt_w_d_out);
|
|
|
|
test.cvt_d_s_in = std::numeric_limits<float>::max();
|
|
test.cvt_d_w_in = std::numeric_limits<int32_t>::max();
|
|
test.cvt_d_l_in = std::numeric_limits<int64_t>::max();
|
|
test.cvt_l_s_in = std::numeric_limits<float>::max();
|
|
test.cvt_l_d_in = std::numeric_limits<double>::max();
|
|
test.cvt_s_d_in = std::numeric_limits<double>::max();
|
|
test.cvt_s_w_in = std::numeric_limits<int32_t>::max();
|
|
test.cvt_s_l_in = std::numeric_limits<int64_t>::max();
|
|
test.cvt_w_s_in = std::numeric_limits<float>::max();
|
|
test.cvt_w_d_in = std::numeric_limits<double>::max();
|
|
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.cvt_d_s_out, static_cast<double>(test.cvt_d_s_in));
|
|
CHECK_EQ(test.cvt_d_w_out, static_cast<double>(test.cvt_d_w_in));
|
|
if ((IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) &&
|
|
IsFp64Mode()) {
|
|
CHECK_EQ(test.cvt_d_l_out, static_cast<double>(test.cvt_d_l_in));
|
|
}
|
|
if (IsFp64Mode()) {
|
|
CHECK_EQ(test.cvt_l_s_out, std::numeric_limits<int64_t>::max());
|
|
CHECK_EQ(test.cvt_l_d_out, std::numeric_limits<int64_t>::max());
|
|
}
|
|
CHECK_EQ(test.cvt_s_d_out, static_cast<float>(test.cvt_s_d_in));
|
|
CHECK_EQ(test.cvt_s_w_out, static_cast<float>(test.cvt_s_w_in));
|
|
if ((IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) &&
|
|
IsFp64Mode()) {
|
|
CHECK_EQ(test.cvt_s_l_out, static_cast<float>(test.cvt_s_l_in));
|
|
}
|
|
CHECK_EQ(test.cvt_w_s_out, std::numeric_limits<int32_t>::max());
|
|
CHECK_EQ(test.cvt_w_d_out, std::numeric_limits<int32_t>::max());
|
|
|
|
|
|
test.cvt_d_s_in = -std::numeric_limits<float>::max(); // lowest()
|
|
test.cvt_d_w_in = std::numeric_limits<int32_t>::min(); // lowest()
|
|
test.cvt_d_l_in = std::numeric_limits<int64_t>::min(); // lowest()
|
|
test.cvt_l_s_in = -std::numeric_limits<float>::max(); // lowest()
|
|
test.cvt_l_d_in = -std::numeric_limits<double>::max(); // lowest()
|
|
test.cvt_s_d_in = -std::numeric_limits<double>::max(); // lowest()
|
|
test.cvt_s_w_in = std::numeric_limits<int32_t>::min(); // lowest()
|
|
test.cvt_s_l_in = std::numeric_limits<int64_t>::min(); // lowest()
|
|
test.cvt_w_s_in = -std::numeric_limits<float>::max(); // lowest()
|
|
test.cvt_w_d_in = -std::numeric_limits<double>::max(); // lowest()
|
|
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.cvt_d_s_out, static_cast<double>(test.cvt_d_s_in));
|
|
CHECK_EQ(test.cvt_d_w_out, static_cast<double>(test.cvt_d_w_in));
|
|
if ((IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) &&
|
|
IsFp64Mode()) {
|
|
CHECK_EQ(test.cvt_d_l_out, static_cast<double>(test.cvt_d_l_in));
|
|
}
|
|
// The returned value when converting from fixed-point to float-point
|
|
// is not consistent between board, simulator and specification
|
|
// in this test case, therefore modifying the test
|
|
if (IsFp64Mode()) {
|
|
CHECK(test.cvt_l_s_out == std::numeric_limits<int64_t>::min() ||
|
|
test.cvt_l_s_out == std::numeric_limits<int64_t>::max());
|
|
CHECK(test.cvt_l_d_out == std::numeric_limits<int64_t>::min() ||
|
|
test.cvt_l_d_out == std::numeric_limits<int64_t>::max());
|
|
}
|
|
CHECK_EQ(test.cvt_s_d_out, static_cast<float>(test.cvt_s_d_in));
|
|
CHECK_EQ(test.cvt_s_w_out, static_cast<float>(test.cvt_s_w_in));
|
|
if ((IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) &&
|
|
IsFp64Mode()) {
|
|
CHECK_EQ(test.cvt_s_l_out, static_cast<float>(test.cvt_s_l_in));
|
|
}
|
|
CHECK(test.cvt_w_s_out == std::numeric_limits<int32_t>::min() ||
|
|
test.cvt_w_s_out == std::numeric_limits<int32_t>::max());
|
|
CHECK(test.cvt_w_d_out == std::numeric_limits<int32_t>::min() ||
|
|
test.cvt_w_d_out == std::numeric_limits<int32_t>::max());
|
|
|
|
|
|
test.cvt_d_s_in = std::numeric_limits<float>::min();
|
|
test.cvt_d_w_in = std::numeric_limits<int32_t>::min();
|
|
test.cvt_d_l_in = std::numeric_limits<int64_t>::min();
|
|
test.cvt_l_s_in = std::numeric_limits<float>::min();
|
|
test.cvt_l_d_in = std::numeric_limits<double>::min();
|
|
test.cvt_s_d_in = std::numeric_limits<double>::min();
|
|
test.cvt_s_w_in = std::numeric_limits<int32_t>::min();
|
|
test.cvt_s_l_in = std::numeric_limits<int64_t>::min();
|
|
test.cvt_w_s_in = std::numeric_limits<float>::min();
|
|
test.cvt_w_d_in = std::numeric_limits<double>::min();
|
|
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.cvt_d_s_out, static_cast<double>(test.cvt_d_s_in));
|
|
CHECK_EQ(test.cvt_d_w_out, static_cast<double>(test.cvt_d_w_in));
|
|
if ((IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) &&
|
|
IsFp64Mode()) {
|
|
CHECK_EQ(test.cvt_d_l_out, static_cast<double>(test.cvt_d_l_in));
|
|
}
|
|
if (IsFp64Mode()) {
|
|
CHECK_EQ(0, test.cvt_l_s_out);
|
|
CHECK_EQ(0, test.cvt_l_d_out);
|
|
}
|
|
CHECK_EQ(test.cvt_s_d_out, static_cast<float>(test.cvt_s_d_in));
|
|
CHECK_EQ(test.cvt_s_w_out, static_cast<float>(test.cvt_s_w_in));
|
|
if ((IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) &&
|
|
IsFp64Mode()) {
|
|
CHECK_EQ(test.cvt_s_l_out, static_cast<float>(test.cvt_s_l_in));
|
|
}
|
|
CHECK_EQ(0, test.cvt_w_s_out);
|
|
CHECK_EQ(0, test.cvt_w_d_out);
|
|
}
|
|
|
|
|
|
TEST(DIV_FMT) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
struct Test {
|
|
double dOp1;
|
|
double dOp2;
|
|
double dRes;
|
|
float fOp1;
|
|
float fOp2;
|
|
float fRes;
|
|
};
|
|
|
|
Test test;
|
|
|
|
// Save FCSR.
|
|
__ cfc1(a1, FCSR);
|
|
// Disable FPU exceptions.
|
|
__ ctc1(zero_reg, FCSR);
|
|
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(Test, dOp1)));
|
|
__ Ldc1(f2, MemOperand(a0, offsetof(Test, dOp2)));
|
|
__ nop();
|
|
__ div_d(f6, f4, f2);
|
|
__ Sdc1(f6, MemOperand(a0, offsetof(Test, dRes)));
|
|
|
|
__ lwc1(f4, MemOperand(a0, offsetof(Test, fOp1)) );
|
|
__ lwc1(f2, MemOperand(a0, offsetof(Test, fOp2)) );
|
|
__ nop();
|
|
__ div_s(f6, f4, f2);
|
|
__ swc1(f6, MemOperand(a0, offsetof(Test, fRes)) );
|
|
|
|
// Restore FCSR.
|
|
__ ctc1(a1, FCSR);
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
|
|
const int test_size = 3;
|
|
|
|
double dOp1[test_size] = {
|
|
5.0,
|
|
DBL_MAX,
|
|
DBL_MAX,
|
|
};
|
|
double dOp2[test_size] = {
|
|
2.0,
|
|
2.0,
|
|
-DBL_MAX,
|
|
};
|
|
double dRes[test_size] = {
|
|
2.5,
|
|
DBL_MAX / 2.0,
|
|
-1.0,
|
|
};
|
|
float fOp1[test_size] = {
|
|
5.0,
|
|
FLT_MAX,
|
|
FLT_MAX,
|
|
};
|
|
float fOp2[test_size] = {
|
|
2.0,
|
|
2.0,
|
|
-FLT_MAX,
|
|
};
|
|
float fRes[test_size] = {
|
|
2.5,
|
|
FLT_MAX / 2.0,
|
|
-1.0,
|
|
};
|
|
|
|
for (int i = 0; i < test_size; i++) {
|
|
test.dOp1 = dOp1[i];
|
|
test.dOp2 = dOp2[i];
|
|
test.fOp1 = fOp1[i];
|
|
test.fOp2 = fOp2[i];
|
|
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK_EQ(test.dRes, dRes[i]);
|
|
CHECK_EQ(test.fRes, fRes[i]);
|
|
}
|
|
|
|
test.dOp1 = DBL_MAX;
|
|
test.dOp2 = -0.0;
|
|
test.fOp1 = FLT_MAX;
|
|
test.fOp2 = -0.0;
|
|
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK(!std::isfinite(test.dRes));
|
|
CHECK(!std::isfinite(test.fRes));
|
|
|
|
test.dOp1 = 0.0;
|
|
test.dOp2 = -0.0;
|
|
test.fOp1 = 0.0;
|
|
test.fOp2 = -0.0;
|
|
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK(std::isnan(test.dRes));
|
|
CHECK(std::isnan(test.fRes));
|
|
|
|
test.dOp1 = std::numeric_limits<double>::quiet_NaN();
|
|
test.dOp2 = -5.0;
|
|
test.fOp1 = std::numeric_limits<float>::quiet_NaN();
|
|
test.fOp2 = -5.0;
|
|
|
|
(f.Call(&test, 0, 0, 0, 0));
|
|
CHECK(std::isnan(test.dRes));
|
|
CHECK(std::isnan(test.fRes));
|
|
}
|
|
|
|
|
|
uint32_t run_align(uint32_t rs_value, uint32_t rt_value, uint8_t bp) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
__ align(v0, a0, a1, bp);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
|
|
auto f = GeneratedCode<F2>::FromCode(*code);
|
|
|
|
uint32_t res =
|
|
reinterpret_cast<uint32_t>(f.Call(rs_value, rt_value, 0, 0, 0));
|
|
|
|
return res;
|
|
}
|
|
|
|
|
|
TEST(r6_align) {
|
|
if (IsMipsArchVariant(kMips32r6)) {
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseAlign {
|
|
uint32_t rs_value;
|
|
uint32_t rt_value;
|
|
uint8_t bp;
|
|
uint32_t expected_res;
|
|
};
|
|
|
|
// clang-format off
|
|
struct TestCaseAlign tc[] = {
|
|
// rs_value, rt_value, bp, expected_res
|
|
{0x11223344, 0xAABBCCDD, 0, 0xAABBCCDD},
|
|
{0x11223344, 0xAABBCCDD, 1, 0xBBCCDD11},
|
|
{0x11223344, 0xAABBCCDD, 2, 0xCCDD1122},
|
|
{0x11223344, 0xAABBCCDD, 3, 0xDD112233},
|
|
};
|
|
// clang-format on
|
|
|
|
size_t nr_test_cases = sizeof(tc) / sizeof(TestCaseAlign);
|
|
for (size_t i = 0; i < nr_test_cases; ++i) {
|
|
CHECK_EQ(tc[i].expected_res, run_align(tc[i].rs_value,
|
|
tc[i].rt_value, tc[i].bp));
|
|
}
|
|
}
|
|
}
|
|
|
|
uint32_t PC; // The program counter.
|
|
|
|
uint32_t run_aluipc(int16_t offset) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
__ aluipc(v0, offset);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
|
|
auto f = GeneratedCode<F2>::FromCode(*code);
|
|
PC = (uint32_t)code->entry(); // Set the program counter.
|
|
|
|
uint32_t res = reinterpret_cast<uint32_t>(f.Call(0, 0, 0, 0, 0));
|
|
|
|
return res;
|
|
}
|
|
|
|
|
|
TEST(r6_aluipc) {
|
|
if (IsMipsArchVariant(kMips32r6)) {
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseAluipc {
|
|
int16_t offset;
|
|
};
|
|
|
|
struct TestCaseAluipc tc[] = {
|
|
// offset
|
|
{ -32768 }, // 0x8000
|
|
{ -1 }, // 0xFFFF
|
|
{ 0 },
|
|
{ 1 },
|
|
{ 32767 }, // 0x7FFF
|
|
};
|
|
|
|
size_t nr_test_cases = sizeof(tc) / sizeof(TestCaseAluipc);
|
|
for (size_t i = 0; i < nr_test_cases; ++i) {
|
|
PC = 0;
|
|
uint32_t res = run_aluipc(tc[i].offset);
|
|
// Now, the program_counter (PC) is set.
|
|
uint32_t expected_res = ~0x0FFFF & (PC + (tc[i].offset << 16));
|
|
CHECK_EQ(expected_res, res);
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
uint32_t run_auipc(int16_t offset) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
__ auipc(v0, offset);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
|
|
auto f = GeneratedCode<F2>::FromCode(*code);
|
|
PC = (uint32_t)code->entry(); // Set the program counter.
|
|
|
|
uint32_t res = reinterpret_cast<uint32_t>(f.Call(0, 0, 0, 0, 0));
|
|
|
|
return res;
|
|
}
|
|
|
|
|
|
TEST(r6_auipc) {
|
|
if (IsMipsArchVariant(kMips32r6)) {
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseAuipc {
|
|
int16_t offset;
|
|
};
|
|
|
|
struct TestCaseAuipc tc[] = {
|
|
// offset
|
|
{ -32768 }, // 0x8000
|
|
{ -1 }, // 0xFFFF
|
|
{ 0 },
|
|
{ 1 },
|
|
{ 32767 }, // 0x7FFF
|
|
};
|
|
|
|
size_t nr_test_cases = sizeof(tc) / sizeof(TestCaseAuipc);
|
|
for (size_t i = 0; i < nr_test_cases; ++i) {
|
|
PC = 0;
|
|
uint32_t res = run_auipc(tc[i].offset);
|
|
// Now, the program_counter (PC) is set.
|
|
uint32_t expected_res = PC + (tc[i].offset << 16);
|
|
CHECK_EQ(expected_res, res);
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
uint32_t run_lwpc(int offset) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
// 256k instructions; 2^8k
|
|
// addiu t7, t0, 0xFFFF; (0x250FFFFF)
|
|
// ...
|
|
// addiu t4, t0, 0x0000; (0x250C0000)
|
|
uint32_t addiu_start_1 = 0x25000000;
|
|
for (int32_t i = 0xFFFFF; i >= 0xC0000; --i) {
|
|
uint32_t addiu_new = addiu_start_1 + i;
|
|
__ dd(addiu_new);
|
|
}
|
|
|
|
__ lwpc(t8, offset); // offset 0; 0xEF080000 (t8 register)
|
|
__ mov(v0, t8);
|
|
|
|
// 256k instructions; 2^8k
|
|
// addiu t0, t0, 0x0000; (0x25080000)
|
|
// ...
|
|
// addiu t3, t0, 0xFFFF; (0x250BFFFF)
|
|
uint32_t addiu_start_2 = 0x25000000;
|
|
for (int32_t i = 0x80000; i <= 0xBFFFF; ++i) {
|
|
uint32_t addiu_new = addiu_start_2 + i;
|
|
__ dd(addiu_new);
|
|
}
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
|
|
auto f = GeneratedCode<F2>::FromCode(*code);
|
|
|
|
uint32_t res = reinterpret_cast<uint32_t>(f.Call(0, 0, 0, 0, 0));
|
|
|
|
return res;
|
|
}
|
|
|
|
|
|
TEST(r6_lwpc) {
|
|
if (IsMipsArchVariant(kMips32r6)) {
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseLwpc {
|
|
int offset;
|
|
uint32_t expected_res;
|
|
};
|
|
|
|
// clang-format off
|
|
struct TestCaseLwpc tc[] = {
|
|
// offset, expected_res
|
|
{ -262144, 0x250FFFFF }, // offset 0x40000
|
|
{ -4, 0x250C0003 },
|
|
{ -1, 0x250C0000 },
|
|
{ 0, 0xEF080000 },
|
|
{ 1, 0x03001025 }, // mov(v0, t8)
|
|
{ 2, 0x25080000 },
|
|
{ 4, 0x25080002 },
|
|
{ 262143, 0x250BFFFD }, // offset 0x3FFFF
|
|
};
|
|
// clang-format on
|
|
|
|
size_t nr_test_cases = sizeof(tc) / sizeof(TestCaseLwpc);
|
|
for (size_t i = 0; i < nr_test_cases; ++i) {
|
|
uint32_t res = run_lwpc(tc[i].offset);
|
|
CHECK_EQ(tc[i].expected_res, res);
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
uint32_t run_jic(int16_t offset) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
Label stop_execution;
|
|
__ push(ra);
|
|
__ li(v0, 0);
|
|
__ li(t1, 0x66);
|
|
|
|
__ addiu(v0, v0, 0x1); // <-- offset = -32
|
|
__ addiu(v0, v0, 0x2);
|
|
__ addiu(v0, v0, 0x10);
|
|
__ addiu(v0, v0, 0x20);
|
|
__ beq(v0, t1, &stop_execution);
|
|
__ nop();
|
|
|
|
__ nal(); // t0 <- program counter
|
|
__ mov(t0, ra);
|
|
__ jic(t0, offset);
|
|
|
|
__ addiu(v0, v0, 0x100);
|
|
__ addiu(v0, v0, 0x200);
|
|
__ addiu(v0, v0, 0x1000);
|
|
__ addiu(v0, v0, 0x2000); // <--- offset = 16
|
|
__ pop(ra);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
__ bind(&stop_execution);
|
|
__ pop(ra);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
|
|
auto f = GeneratedCode<F2>::FromCode(*code);
|
|
|
|
uint32_t res = reinterpret_cast<uint32_t>(f.Call(0, 0, 0, 0, 0));
|
|
|
|
return res;
|
|
}
|
|
|
|
|
|
TEST(r6_jic) {
|
|
if (IsMipsArchVariant(kMips32r6)) {
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseJic {
|
|
// As rt will be used t0 register which will have value of
|
|
// the program counter for the jic instruction.
|
|
int16_t offset;
|
|
uint32_t expected_res;
|
|
};
|
|
|
|
struct TestCaseJic tc[] = {
|
|
// offset, expected_result
|
|
{ 16, 0x2033 },
|
|
{ 4, 0x3333 },
|
|
{ -32, 0x66 },
|
|
};
|
|
|
|
size_t nr_test_cases = sizeof(tc) / sizeof(TestCaseJic);
|
|
for (size_t i = 0; i < nr_test_cases; ++i) {
|
|
uint32_t res = run_jic(tc[i].offset);
|
|
CHECK_EQ(tc[i].expected_res, res);
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
uint64_t run_beqzc(int32_t value, int32_t offset) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
Label stop_execution;
|
|
__ li(v0, 0);
|
|
__ li(t1, 0x66);
|
|
|
|
__ addiu(v0, v0, 0x1); // <-- offset = -32
|
|
__ addiu(v0, v0, 0x2);
|
|
__ addiu(v0, v0, 0x10);
|
|
__ addiu(v0, v0, 0x20);
|
|
__ beq(v0, t1, &stop_execution);
|
|
__ nop();
|
|
|
|
__ beqzc(a0, offset); // BEQZC rs, offset
|
|
|
|
__ addiu(v0, v0, 0x1);
|
|
__ addiu(v0, v0, 0x100);
|
|
__ addiu(v0, v0, 0x200);
|
|
__ addiu(v0, v0, 0x1000);
|
|
__ addiu(v0, v0, 0x2000); // <--- offset = 16
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
__ bind(&stop_execution);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
|
|
auto f = GeneratedCode<F2>::FromCode(*code);
|
|
|
|
uint32_t res = reinterpret_cast<uint32_t>(f.Call(value, 0, 0, 0, 0));
|
|
|
|
return res;
|
|
}
|
|
|
|
|
|
TEST(r6_beqzc) {
|
|
if (IsMipsArchVariant(kMips32r6)) {
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseBeqzc {
|
|
uint32_t value;
|
|
int32_t offset;
|
|
uint32_t expected_res;
|
|
};
|
|
|
|
// clang-format off
|
|
struct TestCaseBeqzc tc[] = {
|
|
// value, offset, expected_res
|
|
{ 0x0, -8, 0x66 },
|
|
{ 0x0, 0, 0x3334 },
|
|
{ 0x0, 1, 0x3333 },
|
|
{ 0xABC, 1, 0x3334 },
|
|
{ 0x0, 4, 0x2033 },
|
|
};
|
|
// clang-format on
|
|
|
|
size_t nr_test_cases = sizeof(tc) / sizeof(TestCaseBeqzc);
|
|
for (size_t i = 0; i < nr_test_cases; ++i) {
|
|
uint32_t res = run_beqzc(tc[i].value, tc[i].offset);
|
|
CHECK_EQ(tc[i].expected_res, res);
|
|
}
|
|
}
|
|
}
|
|
|
|
void load_elements_of_vector(MacroAssembler* assm_ptr,
|
|
const uint64_t elements[], MSARegister w,
|
|
Register t0, Register t1) {
|
|
MacroAssembler& assm = *assm_ptr;
|
|
__ li(t0, static_cast<uint32_t>(elements[0] & 0xFFFFFFFF));
|
|
__ li(t1, static_cast<uint32_t>((elements[0] >> 32) & 0xFFFFFFFF));
|
|
__ insert_w(w, 0, t0);
|
|
__ insert_w(w, 1, t1);
|
|
__ li(t0, static_cast<uint32_t>(elements[1] & 0xFFFFFFFF));
|
|
__ li(t1, static_cast<uint32_t>((elements[1] >> 32) & 0xFFFFFFFF));
|
|
__ insert_w(w, 2, t0);
|
|
__ insert_w(w, 3, t1);
|
|
}
|
|
|
|
inline void store_elements_of_vector(MacroAssembler* assm_ptr, MSARegister w,
|
|
Register a) {
|
|
MacroAssembler& assm = *assm_ptr;
|
|
__ st_d(w, MemOperand(a, 0));
|
|
}
|
|
|
|
union msa_reg_t {
|
|
uint8_t b[16];
|
|
uint16_t h[8];
|
|
uint32_t w[4];
|
|
uint64_t d[2];
|
|
};
|
|
|
|
struct TestCaseMsaBranch {
|
|
uint64_t wt_lo;
|
|
uint64_t wt_hi;
|
|
};
|
|
|
|
template <typename Branch>
|
|
void run_bz_bnz(TestCaseMsaBranch* input, Branch GenerateBranch,
|
|
bool branched) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
CpuFeatureScope fscope(&assm, MIPS_SIMD);
|
|
|
|
struct T {
|
|
uint64_t ws_lo;
|
|
uint64_t ws_hi;
|
|
uint64_t wd_lo;
|
|
uint64_t wd_hi;
|
|
};
|
|
T t = {0x20B9CC4F1A83E0C5, 0xA27E1B5F2F5BB18A, 0x0000000000000000,
|
|
0x0000000000000000};
|
|
msa_reg_t res;
|
|
Label do_not_move_w0_to_w2;
|
|
|
|
load_elements_of_vector(&assm, &t.ws_lo, w0, t0, t1);
|
|
load_elements_of_vector(&assm, &t.wd_lo, w2, t0, t1);
|
|
load_elements_of_vector(&assm, &input->wt_lo, w1, t0, t1);
|
|
GenerateBranch(assm, do_not_move_w0_to_w2);
|
|
__ nop();
|
|
__ move_v(w2, w0);
|
|
|
|
__ bind(&do_not_move_w0_to_w2);
|
|
store_elements_of_vector(&assm, w2, a0);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
#ifdef OBJECT_PRINT
|
|
code->Print(std::cout);
|
|
#endif
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
|
|
(f.Call(&res, 0, 0, 0, 0));
|
|
if (branched) {
|
|
CHECK_EQ(t.wd_lo, res.d[0]);
|
|
CHECK_EQ(t.wd_hi, res.d[1]);
|
|
} else {
|
|
CHECK_EQ(t.ws_lo, res.d[0]);
|
|
CHECK_EQ(t.ws_hi, res.d[1]);
|
|
}
|
|
}
|
|
|
|
TEST(MSA_bz_bnz) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
TestCaseMsaBranch tz_v[] = {
|
|
{0x0, 0x0}, {0xABC, 0x0}, {0x0, 0xABC}, {0xABC, 0xABC}};
|
|
for (unsigned i = 0; i < arraysize(tz_v); ++i) {
|
|
run_bz_bnz(
|
|
&tz_v[i],
|
|
[](MacroAssembler& assm, Label& br_target) { __ bz_v(w1, &br_target); },
|
|
tz_v[i].wt_lo == 0 && tz_v[i].wt_hi == 0);
|
|
}
|
|
|
|
#define TEST_BZ_DF(input_array, lanes, instruction, int_type) \
|
|
for (unsigned i = 0; i < arraysize(input_array); ++i) { \
|
|
int j; \
|
|
int_type* element = reinterpret_cast<int_type*>(&input_array[i]); \
|
|
for (j = 0; j < lanes; ++j) { \
|
|
if (element[j] == 0) { \
|
|
break; \
|
|
} \
|
|
} \
|
|
run_bz_bnz(&input_array[i], \
|
|
[](MacroAssembler& assm, Label& br_target) { \
|
|
__ instruction(w1, &br_target); \
|
|
}, \
|
|
j != lanes); \
|
|
}
|
|
TestCaseMsaBranch tz_b[] = {{0x0, 0x0},
|
|
{0xBC0000, 0x0},
|
|
{0x0, 0xAB000000000000CD},
|
|
{0x123456789ABCDEF0, 0xAAAAAAAAAAAAAAAA}};
|
|
TEST_BZ_DF(tz_b, kMSALanesByte, bz_b, int8_t)
|
|
|
|
TestCaseMsaBranch tz_h[] = {{0x0, 0x0},
|
|
{0xBCDE0000, 0x0},
|
|
{0x0, 0xABCD00000000ABCD},
|
|
{0x123456789ABCDEF0, 0xAAAAAAAAAAAAAAAA}};
|
|
TEST_BZ_DF(tz_h, kMSALanesHalf, bz_h, int16_t)
|
|
|
|
TestCaseMsaBranch tz_w[] = {{0x0, 0x0},
|
|
{0xBCDE123400000000, 0x0},
|
|
{0x0, 0x000000001234ABCD},
|
|
{0x123456789ABCDEF0, 0xAAAAAAAAAAAAAAAA}};
|
|
TEST_BZ_DF(tz_w, kMSALanesWord, bz_w, int32_t)
|
|
|
|
TestCaseMsaBranch tz_d[] = {{0x0, 0x0},
|
|
{0xBCDE0000, 0x0},
|
|
{0x0, 0xABCD00000000ABCD},
|
|
{0x123456789ABCDEF0, 0xAAAAAAAAAAAAAAAA}};
|
|
TEST_BZ_DF(tz_d, kMSALanesDword, bz_d, int64_t)
|
|
#undef TEST_BZ_DF
|
|
|
|
TestCaseMsaBranch tnz_v[] = {
|
|
{0x0, 0x0}, {0xABC, 0x0}, {0x0, 0xABC}, {0xABC, 0xABC}};
|
|
for (unsigned i = 0; i < arraysize(tnz_v); ++i) {
|
|
run_bz_bnz(&tnz_v[i],
|
|
[](MacroAssembler& assm, Label& br_target) {
|
|
__ bnz_v(w1, &br_target);
|
|
},
|
|
tnz_v[i].wt_lo != 0 || tnz_v[i].wt_hi != 0);
|
|
}
|
|
|
|
#define TEST_BNZ_DF(input_array, lanes, instruction, int_type) \
|
|
for (unsigned i = 0; i < arraysize(input_array); ++i) { \
|
|
int j; \
|
|
int_type* element = reinterpret_cast<int_type*>(&input_array[i]); \
|
|
for (j = 0; j < lanes; ++j) { \
|
|
if (element[j] == 0) { \
|
|
break; \
|
|
} \
|
|
} \
|
|
run_bz_bnz(&input_array[i], \
|
|
[](MacroAssembler& assm, Label& br_target) { \
|
|
__ instruction(w1, &br_target); \
|
|
}, \
|
|
j == lanes); \
|
|
}
|
|
TestCaseMsaBranch tnz_b[] = {{0x0, 0x0},
|
|
{0xBC0000, 0x0},
|
|
{0x0, 0xAB000000000000CD},
|
|
{0x123456789ABCDEF0, 0xAAAAAAAAAAAAAAAA}};
|
|
TEST_BNZ_DF(tnz_b, 16, bnz_b, int8_t)
|
|
|
|
TestCaseMsaBranch tnz_h[] = {{0x0, 0x0},
|
|
{0xBCDE0000, 0x0},
|
|
{0x0, 0xABCD00000000ABCD},
|
|
{0x123456789ABCDEF0, 0xAAAAAAAAAAAAAAAA}};
|
|
TEST_BNZ_DF(tnz_h, 8, bnz_h, int16_t)
|
|
|
|
TestCaseMsaBranch tnz_w[] = {{0x0, 0x0},
|
|
{0xBCDE123400000000, 0x0},
|
|
{0x0, 0x000000001234ABCD},
|
|
{0x123456789ABCDEF0, 0xAAAAAAAAAAAAAAAA}};
|
|
TEST_BNZ_DF(tnz_w, 4, bnz_w, int32_t)
|
|
|
|
TestCaseMsaBranch tnz_d[] = {{0x0, 0x0},
|
|
{0xBCDE0000, 0x0},
|
|
{0x0, 0xABCD00000000ABCD},
|
|
{0x123456789ABCDEF0, 0xAAAAAAAAAAAAAAAA}};
|
|
TEST_BNZ_DF(tnz_d, 2, bnz_d, int64_t)
|
|
#undef TEST_BNZ_DF
|
|
}
|
|
|
|
uint32_t run_jialc(int16_t offset) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
Label main_block;
|
|
__ push(ra);
|
|
__ li(v0, 0);
|
|
__ beq(v0, v0, &main_block);
|
|
__ nop();
|
|
|
|
// Block 1
|
|
__ addiu(v0, v0, 0x1); // <-- offset = -40
|
|
__ addiu(v0, v0, 0x2);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
// Block 2
|
|
__ addiu(v0, v0, 0x10); // <-- offset = -24
|
|
__ addiu(v0, v0, 0x20);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
// Block 3 (Main)
|
|
__ bind(&main_block);
|
|
__ nal(); // t0 <- program counter
|
|
__ mov(t0, ra);
|
|
__ jialc(t0, offset);
|
|
__ addiu(v0, v0, 0x4);
|
|
__ pop(ra);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
// Block 4
|
|
__ addiu(v0, v0, 0x100); // <-- offset = 20
|
|
__ addiu(v0, v0, 0x200);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
// Block 5
|
|
__ addiu(v0, v0, 0x1000); // <--- offset = 36
|
|
__ addiu(v0, v0, 0x2000);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
|
|
auto f = GeneratedCode<F2>::FromCode(*code);
|
|
|
|
uint32_t res = reinterpret_cast<uint32_t>(f.Call(0, 0, 0, 0, 0));
|
|
|
|
return res;
|
|
}
|
|
|
|
|
|
TEST(r6_jialc) {
|
|
if (IsMipsArchVariant(kMips32r6)) {
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseJialc {
|
|
int16_t offset;
|
|
uint32_t expected_res;
|
|
};
|
|
|
|
struct TestCaseJialc tc[] = {
|
|
// offset, expected_res
|
|
{ -40, 0x7 },
|
|
{ -24, 0x34 },
|
|
{ 20, 0x304 },
|
|
{ 36, 0x3004 }
|
|
};
|
|
|
|
size_t nr_test_cases = sizeof(tc) / sizeof(TestCaseJialc);
|
|
for (size_t i = 0; i < nr_test_cases; ++i) {
|
|
uint32_t res = run_jialc(tc[i].offset);
|
|
CHECK_EQ(tc[i].expected_res, res);
|
|
}
|
|
}
|
|
}
|
|
|
|
static uint32_t run_addiupc(int32_t imm19) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
__ addiupc(v0, imm19);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
|
|
auto f = GeneratedCode<F2>::FromCode(*code);
|
|
PC = (uint32_t)code->entry(); // Set the program counter.
|
|
|
|
uint32_t rs = reinterpret_cast<uint32_t>(f.Call(imm19, 0, 0, 0, 0));
|
|
|
|
return rs;
|
|
}
|
|
|
|
|
|
TEST(r6_addiupc) {
|
|
if (IsMipsArchVariant(kMips32r6)) {
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseAddiupc {
|
|
int32_t imm19;
|
|
};
|
|
|
|
TestCaseAddiupc tc[] = {
|
|
// imm19
|
|
{-262144}, // 0x40000
|
|
{-1}, // 0x7FFFF
|
|
{0},
|
|
{1}, // 0x00001
|
|
{262143} // 0x3FFFF
|
|
};
|
|
|
|
size_t nr_test_cases = sizeof(tc) / sizeof(TestCaseAddiupc);
|
|
for (size_t i = 0; i < nr_test_cases; ++i) {
|
|
PC = 0;
|
|
uint32_t res = run_addiupc(tc[i].imm19);
|
|
// Now, the program_counter (PC) is set.
|
|
uint32_t expected_res = PC + (tc[i].imm19 << 2);
|
|
CHECK_EQ(expected_res, res);
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
int32_t run_bc(int32_t offset) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
Label continue_1, stop_execution;
|
|
__ push(ra);
|
|
__ li(v0, 0);
|
|
__ li(t8, 0);
|
|
__ li(t9, 2); // A condition for stopping execution.
|
|
|
|
for (int32_t i = -100; i <= -11; ++i) {
|
|
__ addiu(v0, v0, 1);
|
|
}
|
|
|
|
__ addiu(t8, t8, 1); // -10
|
|
|
|
__ beq(t8, t9, &stop_execution); // -9
|
|
__ nop(); // -8
|
|
__ beq(t8, t8, &continue_1); // -7
|
|
__ nop(); // -6
|
|
|
|
__ bind(&stop_execution);
|
|
__ pop(ra); // -5, -4
|
|
__ jr(ra); // -3
|
|
__ nop(); // -2
|
|
|
|
__ bind(&continue_1);
|
|
__ bc(offset); // -1
|
|
|
|
for (int32_t i = 0; i <= 99; ++i) {
|
|
__ addiu(v0, v0, 1);
|
|
}
|
|
|
|
__ pop(ra);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
|
|
auto f = GeneratedCode<F2>::FromCode(*code);
|
|
|
|
int32_t res = reinterpret_cast<int32_t>(f.Call(0, 0, 0, 0, 0));
|
|
|
|
return res;
|
|
}
|
|
|
|
|
|
TEST(r6_bc) {
|
|
if (IsMipsArchVariant(kMips32r6)) {
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseBc {
|
|
int32_t offset;
|
|
int32_t expected_res;
|
|
};
|
|
|
|
struct TestCaseBc tc[] = {
|
|
// offset, expected_result
|
|
{ -100, (abs(-100) - 10) * 2 },
|
|
{ -11, (abs(-100) - 10 + 1) },
|
|
{ 0, (abs(-100) - 10 + 1 + 99) },
|
|
{ 1, (abs(-100) - 10 + 99) },
|
|
{ 99, (abs(-100) - 10 + 1) },
|
|
};
|
|
|
|
size_t nr_test_cases = sizeof(tc) / sizeof(TestCaseBc);
|
|
for (size_t i = 0; i < nr_test_cases; ++i) {
|
|
int32_t res = run_bc(tc[i].offset);
|
|
CHECK_EQ(tc[i].expected_res, res);
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
int32_t run_balc(int32_t offset) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
Label continue_1;
|
|
__ push(ra);
|
|
__ li(v0, 0);
|
|
__ li(t8, 0);
|
|
__ li(t9, 2); // A condition for stopping execution.
|
|
|
|
__ beq(t8, t8, &continue_1);
|
|
__ nop();
|
|
|
|
uint32_t instruction_addiu = 0x24420001; // addiu v0, v0, 1
|
|
for (int32_t i = -117; i <= -57; ++i) {
|
|
__ dd(instruction_addiu);
|
|
}
|
|
__ jr(ra); // -56
|
|
__ nop(); // -55
|
|
|
|
for (int32_t i = -54; i <= -4; ++i) {
|
|
__ dd(instruction_addiu);
|
|
}
|
|
__ jr(ra); // -3
|
|
__ nop(); // -2
|
|
|
|
__ bind(&continue_1);
|
|
__ balc(offset); // -1
|
|
|
|
__ pop(ra); // 0, 1
|
|
__ jr(ra); // 2
|
|
__ nop(); // 3
|
|
|
|
for (int32_t i = 4; i <= 44; ++i) {
|
|
__ dd(instruction_addiu);
|
|
}
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
|
|
auto f = GeneratedCode<F2>::FromCode(*code);
|
|
|
|
int32_t res = reinterpret_cast<int32_t>(f.Call(0, 0, 0, 0, 0));
|
|
|
|
return res;
|
|
}
|
|
|
|
|
|
uint32_t run_aui(uint32_t rs, uint16_t offset) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
__ li(t0, rs);
|
|
__ aui(v0, t0, offset);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
|
|
auto f = GeneratedCode<F2>::FromCode(*code);
|
|
|
|
uint32_t res = reinterpret_cast<uint32_t>(f.Call(0, 0, 0, 0, 0));
|
|
|
|
return res;
|
|
}
|
|
|
|
|
|
TEST(r6_aui) {
|
|
if (IsMipsArchVariant(kMips32r6)) {
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseAui {
|
|
uint32_t rs;
|
|
uint16_t offset;
|
|
uint32_t ref_res;
|
|
};
|
|
|
|
struct TestCaseAui tc[] = {
|
|
// input, offset, result
|
|
{0xFFFEFFFF, 1, 0xFFFFFFFF},
|
|
{0xFFFFFFFF, 0, 0xFFFFFFFF},
|
|
{0, 0xFFFF, 0xFFFF0000},
|
|
{0x0008FFFF, 0xFFF7, 0xFFFFFFFF},
|
|
{32767, 32767, 0x7FFF7FFF},
|
|
// overflow cases
|
|
{0xFFFFFFFF, 0x1, 0x0000FFFF},
|
|
{0xFFFFFFFF, 0xFFFF, 0xFFFEFFFF},
|
|
};
|
|
|
|
size_t nr_test_cases = sizeof(tc) / sizeof(TestCaseAui);
|
|
for (size_t i = 0; i < nr_test_cases; ++i) {
|
|
PC = 0;
|
|
uint32_t res = run_aui(tc[i].rs, tc[i].offset);
|
|
CHECK_EQ(tc[i].ref_res, res);
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
TEST(r6_balc) {
|
|
if (IsMipsArchVariant(kMips32r6)) {
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseBalc {
|
|
int32_t offset;
|
|
int32_t expected_res;
|
|
};
|
|
|
|
struct TestCaseBalc tc[] = {
|
|
// offset, expected_result
|
|
{ -117, 61 },
|
|
{ -54, 51 },
|
|
{ 0, 0 },
|
|
{ 4, 41 },
|
|
};
|
|
|
|
size_t nr_test_cases = sizeof(tc) / sizeof(TestCaseBalc);
|
|
for (size_t i = 0; i < nr_test_cases; ++i) {
|
|
int32_t res = run_balc(tc[i].offset);
|
|
CHECK_EQ(tc[i].expected_res, res);
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
uint32_t run_bal(int16_t offset) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
__ mov(t0, ra);
|
|
__ bal(offset); // Equivalent for "BGEZAL zero_reg, offset".
|
|
__ nop();
|
|
|
|
__ mov(ra, t0);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
__ li(v0, 1);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
|
|
auto f = GeneratedCode<F2>::FromCode(*code);
|
|
|
|
uint32_t res = reinterpret_cast<uint32_t>(f.Call(0, 0, 0, 0, 0));
|
|
|
|
return res;
|
|
}
|
|
|
|
|
|
TEST(bal) {
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseBal {
|
|
int16_t offset;
|
|
uint32_t expected_res;
|
|
};
|
|
|
|
struct TestCaseBal tc[] = {
|
|
// offset, expected_res
|
|
{ 4, 1 },
|
|
};
|
|
|
|
size_t nr_test_cases = sizeof(tc) / sizeof(TestCaseBal);
|
|
for (size_t i = 0; i < nr_test_cases; ++i) {
|
|
CHECK_EQ(tc[i].expected_res, run_bal(tc[i].offset));
|
|
}
|
|
}
|
|
|
|
|
|
TEST(Trampoline) {
|
|
// Private member of Assembler class.
|
|
static const int kMaxBranchOffset = (1 << (18 - 1)) - 1;
|
|
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
Label done;
|
|
size_t nr_calls = kMaxBranchOffset / (2 * kInstrSize) + 2;
|
|
|
|
for (size_t i = 0; i < nr_calls; ++i) {
|
|
__ BranchShort(&done, eq, a0, Operand(a1));
|
|
}
|
|
__ bind(&done);
|
|
__ Ret(USE_DELAY_SLOT);
|
|
__ mov(v0, zero_reg);
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F2>::FromCode(*code);
|
|
|
|
int32_t res = reinterpret_cast<int32_t>(f.Call(42, 42, 0, 0, 0));
|
|
CHECK_EQ(0, res);
|
|
}
|
|
|
|
TEST(Trampoline_with_massive_unbound_labels) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
const int kNumSlots =
|
|
TurboAssembler::kMaxBranchOffset / TurboAssembler::kTrampolineSlotsSize;
|
|
Label labels[kNumSlots];
|
|
|
|
{
|
|
TurboAssembler::BlockTrampolinePoolScope block_trampoline_pool(&assm);
|
|
for (int i = 0; i < kNumSlots; i++) {
|
|
__ Branch(&labels[i]);
|
|
}
|
|
}
|
|
|
|
__ bind(&labels[0]);
|
|
}
|
|
|
|
static void DummyFunction(Object result) {}
|
|
|
|
TEST(Call_with_trampoline) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
int next_buffer_check_ = FLAG_force_long_branches
|
|
? kMaxInt
|
|
: TurboAssembler::kMaxBranchOffset -
|
|
TurboAssembler::kTrampolineSlotsSize * 16;
|
|
|
|
Label done;
|
|
__ Branch(&done);
|
|
next_buffer_check_ -= TurboAssembler::kTrampolineSlotsSize;
|
|
|
|
int num_nops = (next_buffer_check_ - __ pc_offset()) / kInstrSize - 1;
|
|
for (int i = 0; i < num_nops; i++) {
|
|
__ nop();
|
|
}
|
|
|
|
int pc_offset_before = __ pc_offset();
|
|
{
|
|
// There should be a trampoline after this Call
|
|
__ Call(FUNCTION_ADDR(DummyFunction), RelocInfo::RUNTIME_ENTRY);
|
|
}
|
|
int pc_offset_after = __ pc_offset();
|
|
int last_call_pc = __ pc_offset_for_safepoint();
|
|
|
|
// Without trampoline, the Call emits no more than 6 instructions, otherwise
|
|
// more than 6 instructions will be generated.
|
|
int num_instrs = 6;
|
|
// pc_offset_after records the offset after trampoline.
|
|
CHECK_GT(pc_offset_after - pc_offset_before, num_instrs * kInstrSize);
|
|
// last_call_pc records the offset before trampoline.
|
|
CHECK_LE(last_call_pc - pc_offset_before, num_instrs * kInstrSize);
|
|
|
|
__ bind(&done);
|
|
}
|
|
|
|
template <class T>
|
|
struct TestCaseMaddMsub {
|
|
T fr, fs, ft, fd_add, fd_sub;
|
|
};
|
|
|
|
template <typename T, typename F>
|
|
void helper_madd_msub_maddf_msubf(F func) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
T x = std::sqrt(static_cast<T>(2.0));
|
|
T y = std::sqrt(static_cast<T>(3.0));
|
|
T z = std::sqrt(static_cast<T>(5.0));
|
|
T x2 = 11.11, y2 = 22.22, z2 = 33.33;
|
|
TestCaseMaddMsub<T> test_cases[] = {
|
|
{x, y, z, 0.0, 0.0},
|
|
{x, y, -z, 0.0, 0.0},
|
|
{x, -y, z, 0.0, 0.0},
|
|
{x, -y, -z, 0.0, 0.0},
|
|
{-x, y, z, 0.0, 0.0},
|
|
{-x, y, -z, 0.0, 0.0},
|
|
{-x, -y, z, 0.0, 0.0},
|
|
{-x, -y, -z, 0.0, 0.0},
|
|
{-3.14, 0.2345, -123.000056, 0.0, 0.0},
|
|
{7.3, -23.257, -357.1357, 0.0, 0.0},
|
|
{x2, y2, z2, 0.0, 0.0},
|
|
{x2, y2, -z2, 0.0, 0.0},
|
|
{x2, -y2, z2, 0.0, 0.0},
|
|
{x2, -y2, -z2, 0.0, 0.0},
|
|
{-x2, y2, z2, 0.0, 0.0},
|
|
{-x2, y2, -z2, 0.0, 0.0},
|
|
{-x2, -y2, z2, 0.0, 0.0},
|
|
{-x2, -y2, -z2, 0.0, 0.0},
|
|
};
|
|
|
|
if (std::is_same<T, float>::value) {
|
|
__ lwc1(f4, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, fr)));
|
|
__ lwc1(f6, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, fs)));
|
|
__ lwc1(f8, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, ft)));
|
|
__ lwc1(f16, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, fr)));
|
|
} else if (std::is_same<T, double>::value) {
|
|
__ Ldc1(f4, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, fr)));
|
|
__ Ldc1(f6, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, fs)));
|
|
__ Ldc1(f8, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, ft)));
|
|
__ Ldc1(f16, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, fr)));
|
|
} else {
|
|
UNREACHABLE();
|
|
}
|
|
|
|
func(assm);
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
|
|
const size_t kTableLength = sizeof(test_cases) / sizeof(TestCaseMaddMsub<T>);
|
|
TestCaseMaddMsub<T> tc;
|
|
for (size_t i = 0; i < kTableLength; i++) {
|
|
tc.fr = test_cases[i].fr;
|
|
tc.fs = test_cases[i].fs;
|
|
tc.ft = test_cases[i].ft;
|
|
|
|
(f.Call(&tc, 0, 0, 0, 0));
|
|
|
|
T res_add = 0;
|
|
T res_sub = 0;
|
|
if (IsMipsArchVariant(kMips32r2)) {
|
|
res_add = (tc.fs * tc.ft) + tc.fr;
|
|
res_sub = (tc.fs * tc.ft) - tc.fr;
|
|
} else if (IsMipsArchVariant(kMips32r6)) {
|
|
res_add = std::fma(tc.fs, tc.ft, tc.fr);
|
|
res_sub = std::fma(-tc.fs, tc.ft, tc.fr);
|
|
} else {
|
|
UNREACHABLE();
|
|
}
|
|
|
|
CHECK_EQ(tc.fd_add, res_add);
|
|
CHECK_EQ(tc.fd_sub, res_sub);
|
|
}
|
|
}
|
|
|
|
TEST(madd_msub_s) {
|
|
if (!IsMipsArchVariant(kMips32r2)) return;
|
|
helper_madd_msub_maddf_msubf<float>([](MacroAssembler& assm) {
|
|
__ madd_s(f10, f4, f6, f8);
|
|
__ swc1(f10, MemOperand(a0, offsetof(TestCaseMaddMsub<float>, fd_add)));
|
|
__ msub_s(f16, f4, f6, f8);
|
|
__ swc1(f16, MemOperand(a0, offsetof(TestCaseMaddMsub<float>, fd_sub)));
|
|
});
|
|
}
|
|
|
|
TEST(madd_msub_d) {
|
|
if (!IsMipsArchVariant(kMips32r2)) return;
|
|
helper_madd_msub_maddf_msubf<double>([](MacroAssembler& assm) {
|
|
__ madd_d(f10, f4, f6, f8);
|
|
__ Sdc1(f10, MemOperand(a0, offsetof(TestCaseMaddMsub<double>, fd_add)));
|
|
__ msub_d(f16, f4, f6, f8);
|
|
__ Sdc1(f16, MemOperand(a0, offsetof(TestCaseMaddMsub<double>, fd_sub)));
|
|
});
|
|
}
|
|
|
|
TEST(maddf_msubf_s) {
|
|
if (!IsMipsArchVariant(kMips32r6)) return;
|
|
helper_madd_msub_maddf_msubf<float>([](MacroAssembler& assm) {
|
|
__ maddf_s(f4, f6, f8);
|
|
__ swc1(f4, MemOperand(a0, offsetof(TestCaseMaddMsub<float>, fd_add)));
|
|
__ msubf_s(f16, f6, f8);
|
|
__ swc1(f16, MemOperand(a0, offsetof(TestCaseMaddMsub<float>, fd_sub)));
|
|
});
|
|
}
|
|
|
|
TEST(maddf_msubf_d) {
|
|
if (!IsMipsArchVariant(kMips32r6)) return;
|
|
helper_madd_msub_maddf_msubf<double>([](MacroAssembler& assm) {
|
|
__ maddf_d(f4, f6, f8);
|
|
__ Sdc1(f4, MemOperand(a0, offsetof(TestCaseMaddMsub<double>, fd_add)));
|
|
__ msubf_d(f16, f6, f8);
|
|
__ Sdc1(f16, MemOperand(a0, offsetof(TestCaseMaddMsub<double>, fd_sub)));
|
|
});
|
|
}
|
|
|
|
uint32_t run_Subu(uint32_t imm, int32_t num_instr) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
Label code_start;
|
|
__ bind(&code_start);
|
|
__ Subu(v0, zero_reg, imm);
|
|
CHECK_EQ(assm.SizeOfCodeGeneratedSince(&code_start), num_instr * kInstrSize);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F2>::FromCode(*code);
|
|
|
|
uint32_t res = reinterpret_cast<uint32_t>(f.Call(0, 0, 0, 0, 0));
|
|
|
|
return res;
|
|
}
|
|
|
|
TEST(Subu) {
|
|
CcTest::InitializeVM();
|
|
|
|
// Test Subu macro-instruction for min_int16 and max_int16 border cases.
|
|
// For subtracting int16 immediate values we use addiu.
|
|
|
|
struct TestCaseSubu {
|
|
uint32_t imm;
|
|
uint32_t expected_res;
|
|
int32_t num_instr;
|
|
};
|
|
|
|
// We call Subu(v0, zero_reg, imm) to test cases listed below.
|
|
// 0 - imm = expected_res
|
|
struct TestCaseSubu tc[] = {
|
|
// imm, expected_res, num_instr
|
|
{0xFFFF8000, 0x8000, 2}, // min_int16
|
|
// Generates ori + addu
|
|
// We can't have just addiu because -min_int16 > max_int16 so use
|
|
// register. We can load min_int16 to at register with addiu and then
|
|
// subtract at with subu, but now we use ori + addu because -min_int16 can
|
|
// be loaded using ori.
|
|
{0x8000, 0xFFFF8000, 1}, // max_int16 + 1
|
|
// Generates addiu
|
|
// max_int16 + 1 is not int16 but -(max_int16 + 1) is, just use addiu.
|
|
{0xFFFF7FFF, 0x8001, 2}, // min_int16 - 1
|
|
// Generates ori + addu
|
|
// To load this value to at we need two instructions and another one to
|
|
// subtract, lui + ori + subu. But we can load -value to at using just
|
|
// ori and then add at register with addu.
|
|
{0x8001, 0xFFFF7FFF, 2}, // max_int16 + 2
|
|
// Generates ori + subu
|
|
// Not int16 but is uint16, load value to at with ori and subtract with
|
|
// subu.
|
|
{0x00010000, 0xFFFF0000, 2},
|
|
// Generates lui + subu
|
|
// Load value using lui to at and subtract with subu.
|
|
{0x00010001, 0xFFFEFFFF, 3},
|
|
// Generates lui + ori + subu
|
|
// We have to generate three instructions in this case.
|
|
};
|
|
|
|
size_t nr_test_cases = sizeof(tc) / sizeof(TestCaseSubu);
|
|
for (size_t i = 0; i < nr_test_cases; ++i) {
|
|
CHECK_EQ(tc[i].expected_res, run_Subu(tc[i].imm, tc[i].num_instr));
|
|
}
|
|
}
|
|
|
|
TEST(MSA_fill_copy) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
struct T {
|
|
uint32_t u8;
|
|
uint32_t u16;
|
|
uint32_t u32;
|
|
uint32_t s8;
|
|
uint32_t s16;
|
|
uint32_t s32;
|
|
};
|
|
T t;
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
{
|
|
CpuFeatureScope fscope(&assm, MIPS_SIMD);
|
|
|
|
__ li(t0, 0xA512B683);
|
|
|
|
__ fill_b(w0, t0);
|
|
__ fill_h(w2, t0);
|
|
__ fill_w(w4, t0);
|
|
__ copy_u_b(t1, w0, 11);
|
|
__ sw(t1, MemOperand(a0, offsetof(T, u8)));
|
|
__ copy_u_h(t1, w2, 6);
|
|
__ sw(t1, MemOperand(a0, offsetof(T, u16)));
|
|
__ copy_u_w(t1, w4, 3);
|
|
__ sw(t1, MemOperand(a0, offsetof(T, u32)));
|
|
|
|
__ copy_s_b(t1, w0, 8);
|
|
__ sw(t1, MemOperand(a0, offsetof(T, s8)));
|
|
__ copy_s_h(t1, w2, 5);
|
|
__ sw(t1, MemOperand(a0, offsetof(T, s16)));
|
|
__ copy_s_w(t1, w4, 1);
|
|
__ sw(t1, MemOperand(a0, offsetof(T, s32)));
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
}
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
#ifdef OBJECT_PRINT
|
|
code->Print(std::cout);
|
|
#endif
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
|
|
f.Call(&t, 0, 0, 0, 0);
|
|
|
|
CHECK_EQ(0x83u, t.u8);
|
|
CHECK_EQ(0xB683u, t.u16);
|
|
CHECK_EQ(0xA512B683u, t.u32);
|
|
CHECK_EQ(0xFFFFFF83u, t.s8);
|
|
CHECK_EQ(0xFFFFB683u, t.s16);
|
|
CHECK_EQ(0xA512B683u, t.s32);
|
|
}
|
|
|
|
TEST(MSA_fill_copy_2) {
|
|
// Similar to MSA_fill_copy test, but also check overlaping between MSA and
|
|
// FPU registers with same numbers
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
struct T {
|
|
uint32_t w0;
|
|
uint32_t w1;
|
|
uint32_t w2;
|
|
uint32_t w3;
|
|
};
|
|
T t[2];
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
{
|
|
CpuFeatureScope fscope(&assm, MIPS_SIMD);
|
|
|
|
__ li(t0, 0xAAAAAAAA);
|
|
__ li(t1, 0x55555555);
|
|
|
|
__ fill_w(w0, t0);
|
|
__ fill_w(w2, t0);
|
|
|
|
__ FmoveLow(f0, t1);
|
|
__ FmoveHigh(f2, t1);
|
|
|
|
#define STORE_MSA_REG(w_reg, base, scratch) \
|
|
__ copy_u_w(scratch, w_reg, 0); \
|
|
__ sw(scratch, MemOperand(base, offsetof(T, w0))); \
|
|
__ copy_u_w(scratch, w_reg, 1); \
|
|
__ sw(scratch, MemOperand(base, offsetof(T, w1))); \
|
|
__ copy_u_w(scratch, w_reg, 2); \
|
|
__ sw(scratch, MemOperand(base, offsetof(T, w2))); \
|
|
__ copy_u_w(scratch, w_reg, 3); \
|
|
__ sw(scratch, MemOperand(base, offsetof(T, w3)));
|
|
|
|
STORE_MSA_REG(w0, a0, t2)
|
|
STORE_MSA_REG(w2, a1, t2)
|
|
#undef STORE_MSA_REG
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
}
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
#ifdef OBJECT_PRINT
|
|
code->Print(std::cout);
|
|
#endif
|
|
auto f = GeneratedCode<F4>::FromCode(*code);
|
|
|
|
f.Call(&t[0], &t[1], 0, 0, 0);
|
|
|
|
CHECK_EQ(0x55555555, t[0].w0);
|
|
CHECK_EQ(0xAAAAAAAA, t[0].w1);
|
|
CHECK_EQ(0xAAAAAAAA, t[0].w2);
|
|
CHECK_EQ(0xAAAAAAAA, t[0].w3);
|
|
CHECK_EQ(0xAAAAAAAA, t[1].w0);
|
|
CHECK_EQ(0x55555555, t[1].w1);
|
|
CHECK_EQ(0xAAAAAAAA, t[1].w2);
|
|
CHECK_EQ(0xAAAAAAAA, t[1].w3);
|
|
}
|
|
|
|
TEST(MSA_fill_copy_3) {
|
|
// Similar to MSA_fill_copy test, but also check overlaping between MSA and
|
|
// FPU registers with same numbers
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
struct T {
|
|
uint64_t d0;
|
|
uint64_t d1;
|
|
};
|
|
T t[2];
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
{
|
|
CpuFeatureScope fscope(&assm, MIPS_SIMD);
|
|
|
|
__ li(t0, 0xAAAAAAAA);
|
|
__ li(t1, 0x55555555);
|
|
|
|
__ Move(f0, t0, t0);
|
|
__ Move(f2, t0, t0);
|
|
|
|
__ fill_w(w0, t1);
|
|
__ fill_w(w2, t1);
|
|
|
|
__ Sdc1(f0, MemOperand(a0, offsetof(T, d0)));
|
|
__ Sdc1(f2, MemOperand(a1, offsetof(T, d0)));
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
}
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
#ifdef OBJECT_PRINT
|
|
code->Print(std::cout);
|
|
#endif
|
|
auto f = GeneratedCode<F4>::FromCode(*code);
|
|
|
|
f.Call(&t[0], &t[1], 0, 0, 0);
|
|
|
|
CHECK_EQ(0x5555555555555555, t[0].d0);
|
|
CHECK_EQ(0x5555555555555555, t[1].d0);
|
|
}
|
|
|
|
template <typename T>
|
|
void run_msa_insert(int32_t rs_value, int n, msa_reg_t* w) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
CpuFeatureScope fscope(&assm, MIPS_SIMD);
|
|
|
|
__ li(t0, -1);
|
|
__ li(t1, rs_value);
|
|
__ fill_w(w0, t0);
|
|
|
|
if (std::is_same<T, int8_t>::value) {
|
|
DCHECK_LT(n, 16);
|
|
__ insert_b(w0, n, t1);
|
|
} else if (std::is_same<T, int16_t>::value) {
|
|
DCHECK_LT(n, 8);
|
|
__ insert_h(w0, n, t1);
|
|
} else if (std::is_same<T, int32_t>::value) {
|
|
DCHECK_LT(n, 4);
|
|
__ insert_w(w0, n, t1);
|
|
} else {
|
|
UNREACHABLE();
|
|
}
|
|
|
|
store_elements_of_vector(&assm, w0, a0);
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
#ifdef OBJECT_PRINT
|
|
code->Print(std::cout);
|
|
#endif
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
|
|
(f.Call(w, 0, 0, 0, 0));
|
|
}
|
|
|
|
TEST(MSA_insert) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseInsert {
|
|
uint32_t input;
|
|
int n;
|
|
uint64_t exp_res_lo;
|
|
uint64_t exp_res_hi;
|
|
};
|
|
|
|
struct TestCaseInsert tc_b[] = {
|
|
// input, n, exp_res_lo, exp_res_hi
|
|
{0xA2, 13, 0xFFFFFFFFFFFFFFFFu, 0xFFFFA2FFFFFFFFFFu},
|
|
{0x73, 10, 0xFFFFFFFFFFFFFFFFu, 0xFFFFFFFFFF73FFFFu},
|
|
{0x3494, 5, 0xFFFF94FFFFFFFFFFu, 0xFFFFFFFFFFFFFFFFu},
|
|
{0xA6B8, 1, 0xFFFFFFFFFFFFB8FFu, 0xFFFFFFFFFFFFFFFFu}};
|
|
|
|
for (size_t i = 0; i < sizeof(tc_b) / sizeof(TestCaseInsert); ++i) {
|
|
msa_reg_t res;
|
|
run_msa_insert<int8_t>(tc_b[i].input, tc_b[i].n, &res);
|
|
CHECK_EQ(tc_b[i].exp_res_lo, res.d[0]);
|
|
CHECK_EQ(tc_b[i].exp_res_hi, res.d[1]);
|
|
}
|
|
|
|
struct TestCaseInsert tc_h[] = {
|
|
// input, n, exp_res_lo, exp_res_hi
|
|
{0x85A2, 7, 0xFFFFFFFFFFFFFFFFu, 0x85A2FFFFFFFFFFFFu},
|
|
{0xE873, 5, 0xFFFFFFFFFFFFFFFFu, 0xFFFFFFFFE873FFFFu},
|
|
{0x3494, 3, 0x3494FFFFFFFFFFFFu, 0xFFFFFFFFFFFFFFFFu},
|
|
{0xA6B8, 1, 0xFFFFFFFFA6B8FFFFu, 0xFFFFFFFFFFFFFFFFu}};
|
|
|
|
for (size_t i = 0; i < sizeof(tc_h) / sizeof(TestCaseInsert); ++i) {
|
|
msa_reg_t res;
|
|
run_msa_insert<int16_t>(tc_h[i].input, tc_h[i].n, &res);
|
|
CHECK_EQ(tc_h[i].exp_res_lo, res.d[0]);
|
|
CHECK_EQ(tc_h[i].exp_res_hi, res.d[1]);
|
|
}
|
|
|
|
struct TestCaseInsert tc_w[] = {
|
|
// input, n, exp_res_lo, exp_res_hi
|
|
{0xD2F085A2u, 3, 0xFFFFFFFFFFFFFFFFu, 0xD2F085A2FFFFFFFFu},
|
|
{0x4567E873u, 2, 0xFFFFFFFFFFFFFFFFu, 0xFFFFFFFF4567E873u},
|
|
{0xACDB3494u, 1, 0xACDB3494FFFFFFFFu, 0xFFFFFFFFFFFFFFFFu},
|
|
{0x89ABA6B8u, 0, 0xFFFFFFFF89ABA6B8u, 0xFFFFFFFFFFFFFFFFu}};
|
|
|
|
for (size_t i = 0; i < sizeof(tc_w) / sizeof(TestCaseInsert); ++i) {
|
|
msa_reg_t res;
|
|
run_msa_insert<int32_t>(tc_w[i].input, tc_w[i].n, &res);
|
|
CHECK_EQ(tc_w[i].exp_res_lo, res.d[0]);
|
|
CHECK_EQ(tc_w[i].exp_res_hi, res.d[1]);
|
|
}
|
|
}
|
|
|
|
TEST(MSA_move_v) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
struct T {
|
|
uint64_t ws_lo;
|
|
uint64_t ws_hi;
|
|
uint64_t wd_lo;
|
|
uint64_t wd_hi;
|
|
};
|
|
T t[] = {{0x20B9CC4F1A83E0C5, 0xA27E1B5F2F5BB18A, 0x1E86678B52F8E1FF,
|
|
0x706E51290AC76FB9},
|
|
{0x4414AED7883FFD18, 0x047D183A06B67016, 0x4EF258CF8D822870,
|
|
0x2686B73484C2E843},
|
|
{0xD38FF9D048884FFC, 0x6DC63A57C0943CA7, 0x8520CA2F3E97C426,
|
|
0xA9913868FB819C59}};
|
|
|
|
for (unsigned i = 0; i < arraysize(t); ++i) {
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
CpuFeatureScope fscope(&assm, MIPS_SIMD);
|
|
|
|
load_elements_of_vector(&assm, &t[i].ws_lo, w0, t0, t1);
|
|
load_elements_of_vector(&assm, &t[i].wd_lo, w2, t0, t1);
|
|
__ move_v(w2, w0);
|
|
store_elements_of_vector(&assm, w2, a0);
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
#ifdef OBJECT_PRINT
|
|
code->Print(std::cout);
|
|
#endif
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
(f.Call(&t[i].wd_lo, 0, 0, 0, 0));
|
|
CHECK_EQ(t[i].ws_lo, t[i].wd_lo);
|
|
CHECK_EQ(t[i].ws_hi, t[i].wd_hi);
|
|
}
|
|
}
|
|
|
|
template <typename ExpectFunc, typename OperFunc>
|
|
void run_msa_sldi(OperFunc GenerateOperation,
|
|
ExpectFunc GenerateExpectedResult) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
struct T {
|
|
uint64_t ws_lo;
|
|
uint64_t ws_hi;
|
|
uint64_t wd_lo;
|
|
uint64_t wd_hi;
|
|
};
|
|
T t[] = {{0x20B9CC4F1A83E0C5, 0xA27E1B5F2F5BB18A, 0x1E86678B52F8E1FF,
|
|
0x706E51290AC76FB9},
|
|
{0x4414AED7883FFD18, 0x047D183A06B67016, 0x4EF258CF8D822870,
|
|
0x2686B73484C2E843},
|
|
{0xD38FF9D048884FFC, 0x6DC63A57C0943CA7, 0x8520CA2F3E97C426,
|
|
0xA9913868FB819C59}};
|
|
uint64_t res[2];
|
|
|
|
for (unsigned i = 0; i < arraysize(t); ++i) {
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
CpuFeatureScope fscope(&assm, MIPS_SIMD);
|
|
load_elements_of_vector(&assm, &t[i].ws_lo, w0, t0, t1);
|
|
load_elements_of_vector(&assm, &t[i].wd_lo, w2, t0, t1);
|
|
GenerateOperation(assm);
|
|
store_elements_of_vector(&assm, w2, a0);
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
#ifdef OBJECT_PRINT
|
|
code->Print(std::cout);
|
|
#endif
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
(f.Call(&res[0], 0, 0, 0, 0));
|
|
GenerateExpectedResult(reinterpret_cast<uint8_t*>(&t[i].ws_lo),
|
|
reinterpret_cast<uint8_t*>(&t[i].wd_lo));
|
|
CHECK_EQ(res[0], t[i].wd_lo);
|
|
CHECK_EQ(res[1], t[i].wd_hi);
|
|
}
|
|
}
|
|
|
|
TEST(MSA_sldi) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
CcTest::InitializeVM();
|
|
|
|
#define SLDI_DF(s, k) \
|
|
uint8_t v[32]; \
|
|
for (unsigned i = 0; i < s; i++) { \
|
|
v[i] = ws[s * k + i]; \
|
|
v[i + s] = wd[s * k + i]; \
|
|
} \
|
|
for (unsigned i = 0; i < s; i++) { \
|
|
wd[s * k + i] = v[i + n]; \
|
|
}
|
|
|
|
for (int n = 0; n < 16; ++n) {
|
|
run_msa_sldi([n](MacroAssembler& assm) { __ sldi_b(w2, w0, n); },
|
|
[n](uint8_t* ws, uint8_t* wd) {
|
|
SLDI_DF(kMSARegSize / sizeof(int8_t) / kBitsPerByte, 0)
|
|
});
|
|
}
|
|
|
|
for (int n = 0; n < 8; ++n) {
|
|
run_msa_sldi([n](MacroAssembler& assm) { __ sldi_h(w2, w0, n); },
|
|
[n](uint8_t* ws, uint8_t* wd) {
|
|
for (int k = 0; k < 2; ++k) {
|
|
SLDI_DF(kMSARegSize / sizeof(int16_t) / kBitsPerByte, k)
|
|
}
|
|
});
|
|
}
|
|
|
|
for (int n = 0; n < 4; ++n) {
|
|
run_msa_sldi([n](MacroAssembler& assm) { __ sldi_w(w2, w0, n); },
|
|
[n](uint8_t* ws, uint8_t* wd) {
|
|
for (int k = 0; k < 4; ++k) {
|
|
SLDI_DF(kMSARegSize / sizeof(int32_t) / kBitsPerByte, k)
|
|
}
|
|
});
|
|
}
|
|
|
|
for (int n = 0; n < 2; ++n) {
|
|
run_msa_sldi([n](MacroAssembler& assm) { __ sldi_d(w2, w0, n); },
|
|
[n](uint8_t* ws, uint8_t* wd) {
|
|
for (int k = 0; k < 8; ++k) {
|
|
SLDI_DF(kMSARegSize / sizeof(int64_t) / kBitsPerByte, k)
|
|
}
|
|
});
|
|
}
|
|
#undef SLDI_DF
|
|
}
|
|
|
|
void run_msa_ctc_cfc(uint32_t value) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
CpuFeatureScope fscope(&assm, MIPS_SIMD);
|
|
|
|
MSAControlRegister msareg = {kMSACSRRegister};
|
|
__ li(t0, value);
|
|
__ li(t2, 0);
|
|
__ cfcmsa(t1, msareg);
|
|
__ ctcmsa(msareg, t0);
|
|
__ cfcmsa(t2, msareg);
|
|
__ ctcmsa(msareg, t1);
|
|
__ sw(t2, MemOperand(a0, 0));
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
#ifdef OBJECT_PRINT
|
|
code->Print(std::cout);
|
|
#endif
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
|
|
uint32_t res;
|
|
(f.Call(&res, 0, 0, 0, 0));
|
|
|
|
CHECK_EQ(value & 0x0167FFFF, res);
|
|
}
|
|
|
|
TEST(MSA_cfc_ctc) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
const uint32_t mask_without_cause = 0xFF9C0FFF;
|
|
const uint32_t mask_always_zero = 0x0167FFFF;
|
|
const uint32_t mask_enables = 0x00000F80;
|
|
uint32_t test_case[] = {0x2D5EDE31, 0x07955425, 0x15B7DBE3, 0x2BF8BC37,
|
|
0xE6AAE923, 0x24D0F68D, 0x41AFA84C, 0x2D6BF64F,
|
|
0x925014BD, 0x4DBA7E61};
|
|
for (unsigned i = 0; i < arraysize(test_case); i++) {
|
|
// Setting enable bits and corresponding cause bits could result in
|
|
// exception raised and this prevents that from happening
|
|
test_case[i] = (~test_case[i] & mask_enables) << 5 |
|
|
(test_case[i] & mask_without_cause);
|
|
run_msa_ctc_cfc(test_case[i] & mask_always_zero);
|
|
}
|
|
}
|
|
|
|
struct ExpResShf {
|
|
uint8_t i8;
|
|
uint64_t lo;
|
|
uint64_t hi;
|
|
};
|
|
|
|
void run_msa_i8(SecondaryField opcode, uint64_t ws_lo, uint64_t ws_hi,
|
|
uint8_t i8) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
CpuFeatureScope fscope(&assm, MIPS_SIMD);
|
|
msa_reg_t res;
|
|
uint64_t wd_lo = 0xF35862E13E38F8B0;
|
|
uint64_t wd_hi = 0x4F41FFDEF2BFE636;
|
|
|
|
#define LOAD_W_REG(lo, hi, w_reg) \
|
|
__ li(t0, static_cast<uint32_t>(lo & 0xFFFFFFFF)); \
|
|
__ li(t1, static_cast<uint32_t>((lo >> 32) & 0xFFFFFFFF)); \
|
|
__ insert_w(w_reg, 0, t0); \
|
|
__ insert_w(w_reg, 1, t1); \
|
|
__ li(t0, static_cast<uint32_t>(hi & 0xFFFFFFFF)); \
|
|
__ li(t1, static_cast<uint32_t>((hi >> 32) & 0xFFFFFFFF)); \
|
|
__ insert_w(w_reg, 2, t0); \
|
|
__ insert_w(w_reg, 3, t1);
|
|
|
|
LOAD_W_REG(ws_lo, ws_hi, w0)
|
|
|
|
switch (opcode) {
|
|
case ANDI_B:
|
|
__ andi_b(w2, w0, i8);
|
|
break;
|
|
case ORI_B:
|
|
__ ori_b(w2, w0, i8);
|
|
break;
|
|
case NORI_B:
|
|
__ nori_b(w2, w0, i8);
|
|
break;
|
|
case XORI_B:
|
|
__ xori_b(w2, w0, i8);
|
|
break;
|
|
case BMNZI_B:
|
|
LOAD_W_REG(wd_lo, wd_hi, w2);
|
|
__ bmnzi_b(w2, w0, i8);
|
|
break;
|
|
case BMZI_B:
|
|
LOAD_W_REG(wd_lo, wd_hi, w2);
|
|
__ bmzi_b(w2, w0, i8);
|
|
break;
|
|
case BSELI_B:
|
|
LOAD_W_REG(wd_lo, wd_hi, w2);
|
|
__ bseli_b(w2, w0, i8);
|
|
break;
|
|
case SHF_B:
|
|
__ shf_b(w2, w0, i8);
|
|
break;
|
|
case SHF_H:
|
|
__ shf_h(w2, w0, i8);
|
|
break;
|
|
case SHF_W:
|
|
__ shf_w(w2, w0, i8);
|
|
break;
|
|
default:
|
|
UNREACHABLE();
|
|
}
|
|
|
|
store_elements_of_vector(&assm, w2, a0);
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
#undef LOAD_W_REG
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
#ifdef OBJECT_PRINT
|
|
code->Print(std::cout);
|
|
#endif
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
|
|
(f.Call(&res, 0, 0, 0, 0));
|
|
|
|
uint64_t mask = i8 * 0x0101010101010101ull;
|
|
switch (opcode) {
|
|
case ANDI_B:
|
|
CHECK_EQ(ws_lo & mask, res.d[0]);
|
|
CHECK_EQ(ws_hi & mask, res.d[1]);
|
|
break;
|
|
case ORI_B:
|
|
CHECK_EQ(ws_lo | mask, res.d[0]);
|
|
CHECK_EQ(ws_hi | mask, res.d[1]);
|
|
break;
|
|
case NORI_B:
|
|
CHECK_EQ(~(ws_lo | mask), res.d[0]);
|
|
CHECK_EQ(~(ws_hi | mask), res.d[1]);
|
|
break;
|
|
case XORI_B:
|
|
CHECK_EQ(ws_lo ^ mask, res.d[0]);
|
|
CHECK_EQ(ws_hi ^ mask, res.d[1]);
|
|
break;
|
|
case BMNZI_B:
|
|
CHECK_EQ((ws_lo & mask) | (wd_lo & ~mask), res.d[0]);
|
|
CHECK_EQ((ws_hi & mask) | (wd_hi & ~mask), res.d[1]);
|
|
break;
|
|
case BMZI_B:
|
|
CHECK_EQ((ws_lo & ~mask) | (wd_lo & mask), res.d[0]);
|
|
CHECK_EQ((ws_hi & ~mask) | (wd_hi & mask), res.d[1]);
|
|
break;
|
|
case BSELI_B:
|
|
CHECK_EQ((ws_lo & ~wd_lo) | (mask & wd_lo), res.d[0]);
|
|
CHECK_EQ((ws_hi & ~wd_hi) | (mask & wd_hi), res.d[1]);
|
|
break;
|
|
case SHF_B: {
|
|
struct ExpResShf exp_b[] = {
|
|
// i8, exp_lo, exp_hi
|
|
{0xFFu, 0x11111111B9B9B9B9, 0xF7F7F7F7C8C8C8C8},
|
|
{0x0u, 0x62626262DFDFDFDF, 0xD6D6D6D6C8C8C8C8},
|
|
{0xE4u, 0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636},
|
|
{0x1Bu, 0x1B756911C3D9A7B9, 0xAE94A5F79C8AEFC8},
|
|
{0xB1u, 0x662B6253E8C4DF12, 0x0D3AD6803F8BC88B},
|
|
{0x4Eu, 0x62E1F358F8B03E38, 0xFFDE4F41E636F2BF},
|
|
{0x27u, 0x1B697511C3A7D9B9, 0xAEA594F79CEF8AC8}};
|
|
for (size_t i = 0; i < sizeof(exp_b) / sizeof(ExpResShf); ++i) {
|
|
if (exp_b[i].i8 == i8) {
|
|
CHECK_EQ(exp_b[i].lo, res.d[0]);
|
|
CHECK_EQ(exp_b[i].hi, res.d[1]);
|
|
}
|
|
}
|
|
} break;
|
|
case SHF_H: {
|
|
struct ExpResShf exp_h[] = {
|
|
// i8, exp_lo, exp_hi
|
|
{0xFFu, 0x1169116911691169, 0xF7A5F7A5F7A5F7A5},
|
|
{0x0u, 0x12DF12DF12DF12DF, 0x8BC88BC88BC88BC8},
|
|
{0xE4u, 0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636},
|
|
{0x1Bu, 0xD9C3B9A7751B1169, 0x8A9CC8EF94AEF7A5},
|
|
{0xB1u, 0x53622B6612DFC4E8, 0x80D63A0D8BC88B3F},
|
|
{0x4Eu, 0x3E38F8B0F35862E1, 0xF2BFE6364F41FFDE},
|
|
{0x27u, 0xD9C3751BB9A71169, 0x8A9C94AEC8EFF7A5}};
|
|
for (size_t i = 0; i < sizeof(exp_h) / sizeof(ExpResShf); ++i) {
|
|
if (exp_h[i].i8 == i8) {
|
|
CHECK_EQ(exp_h[i].lo, res.d[0]);
|
|
CHECK_EQ(exp_h[i].hi, res.d[1]);
|
|
}
|
|
}
|
|
} break;
|
|
case SHF_W: {
|
|
struct ExpResShf exp_w[] = {
|
|
// i8, exp_lo, exp_hi
|
|
{0xFFu, 0xF7A594AEF7A594AE, 0xF7A594AEF7A594AE},
|
|
{0x0u, 0xC4E812DFC4E812DF, 0xC4E812DFC4E812DF},
|
|
{0xE4u, 0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636},
|
|
{0x1Bu, 0xC8EF8A9CF7A594AE, 0xB9A7D9C31169751B},
|
|
{0xB1u, 0xC4E812DF2B665362, 0x8B3F8BC83A0D80D6},
|
|
{0x4Eu, 0x4F41FFDEF2BFE636, 0xF35862E13E38F8B0},
|
|
{0x27u, 0x1169751BF7A594AE, 0xB9A7D9C3C8EF8A9C}};
|
|
for (size_t i = 0; i < sizeof(exp_w) / sizeof(ExpResShf); ++i) {
|
|
if (exp_w[i].i8 == i8) {
|
|
CHECK_EQ(exp_w[i].lo, res.d[0]);
|
|
CHECK_EQ(exp_w[i].hi, res.d[1]);
|
|
}
|
|
}
|
|
} break;
|
|
default:
|
|
UNREACHABLE();
|
|
}
|
|
}
|
|
|
|
struct TestCaseMsaI8 {
|
|
uint64_t input_lo;
|
|
uint64_t input_hi;
|
|
uint8_t i8;
|
|
};
|
|
|
|
TEST(MSA_andi_ori_nori_xori) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseMsaI8 tc[] = {// input_lo, input_hi, i8
|
|
{0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0xFFu},
|
|
{0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0x0u},
|
|
{0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0x3Bu},
|
|
{0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0xD9u}};
|
|
|
|
for (size_t i = 0; i < sizeof(tc) / sizeof(TestCaseMsaI8); ++i) {
|
|
run_msa_i8(ANDI_B, tc[i].input_lo, tc[i].input_hi, tc[i].i8);
|
|
run_msa_i8(ORI_B, tc[i].input_lo, tc[i].input_hi, tc[i].i8);
|
|
run_msa_i8(NORI_B, tc[i].input_lo, tc[i].input_hi, tc[i].i8);
|
|
run_msa_i8(XORI_B, tc[i].input_lo, tc[i].input_hi, tc[i].i8);
|
|
}
|
|
}
|
|
|
|
TEST(MSA_bmnzi_bmzi_bseli) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseMsaI8 tc[] = {// input_lo, input_hi, i8
|
|
{0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0xFFu},
|
|
{0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0x0u},
|
|
{0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0x3Bu},
|
|
{0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0xD9u}};
|
|
|
|
for (size_t i = 0; i < sizeof(tc) / sizeof(TestCaseMsaI8); ++i) {
|
|
run_msa_i8(BMNZI_B, tc[i].input_lo, tc[i].input_hi, tc[i].i8);
|
|
run_msa_i8(BMZI_B, tc[i].input_lo, tc[i].input_hi, tc[i].i8);
|
|
run_msa_i8(BSELI_B, tc[i].input_lo, tc[i].input_hi, tc[i].i8);
|
|
}
|
|
}
|
|
|
|
TEST(MSA_shf) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseMsaI8 tc[] = {
|
|
// input_lo, input_hi, i8
|
|
{0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0xFFu}, // 3333
|
|
{0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0x0u}, // 0000
|
|
{0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636, 0xE4u}, // 3210
|
|
{0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0x1Bu}, // 0123
|
|
{0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0xB1u}, // 2301
|
|
{0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636, 0x4Eu}, // 1032
|
|
{0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0x27u} // 0213
|
|
};
|
|
|
|
for (size_t i = 0; i < sizeof(tc) / sizeof(TestCaseMsaI8); ++i) {
|
|
run_msa_i8(SHF_B, tc[i].input_lo, tc[i].input_hi, tc[i].i8);
|
|
run_msa_i8(SHF_H, tc[i].input_lo, tc[i].input_hi, tc[i].i8);
|
|
run_msa_i8(SHF_W, tc[i].input_lo, tc[i].input_hi, tc[i].i8);
|
|
}
|
|
}
|
|
|
|
uint32_t run_Ins(uint32_t imm, uint32_t source, uint16_t pos, uint16_t size) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
__ li(v0, imm);
|
|
__ li(t0, source);
|
|
__ Ins(v0, t0, pos, size);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F2>::FromCode(*code);
|
|
|
|
uint32_t res = reinterpret_cast<uint32_t>(f.Call(0, 0, 0, 0, 0));
|
|
|
|
return res;
|
|
}
|
|
|
|
TEST(Ins) {
|
|
CcTest::InitializeVM();
|
|
|
|
// run_Ins(rt_value, rs_value, pos, size), expected_result
|
|
CHECK_EQ(run_Ins(0x55555555, 0xABCDEF01, 31, 1), 0xD5555555);
|
|
CHECK_EQ(run_Ins(0x55555555, 0xABCDEF02, 30, 2), 0x95555555);
|
|
CHECK_EQ(run_Ins(0x01234567, 0xFABCDEFF, 0, 32), 0xFABCDEFF);
|
|
|
|
// Results with positive sign.
|
|
CHECK_EQ(run_Ins(0x55555550, 0x80000001, 0, 1), 0x55555551);
|
|
CHECK_EQ(run_Ins(0x55555555, 0x40000001, 0, 32), 0x40000001);
|
|
CHECK_EQ(run_Ins(0x55555555, 0x20000001, 1, 31), 0x40000003);
|
|
CHECK_EQ(run_Ins(0x55555555, 0x80700001, 8, 24), 0x70000155);
|
|
CHECK_EQ(run_Ins(0x55555555, 0x80007001, 16, 16), 0x70015555);
|
|
CHECK_EQ(run_Ins(0x55555555, 0x80000071, 24, 8), 0x71555555);
|
|
CHECK_EQ(run_Ins(0x75555555, 0x40000000, 31, 1), 0x75555555);
|
|
|
|
// Results with negative sign.
|
|
CHECK_EQ(run_Ins(0x85555550, 0x80000001, 0, 1), 0x85555551);
|
|
CHECK_EQ(run_Ins(0x55555555, 0x80000001, 0, 32), 0x80000001);
|
|
CHECK_EQ(run_Ins(0x55555555, 0x40000001, 1, 31), 0x80000003);
|
|
CHECK_EQ(run_Ins(0x55555555, 0x80800001, 8, 24), 0x80000155);
|
|
CHECK_EQ(run_Ins(0x55555555, 0x80008001, 16, 16), 0x80015555);
|
|
CHECK_EQ(run_Ins(0x55555555, 0x80000081, 24, 8), 0x81555555);
|
|
CHECK_EQ(run_Ins(0x75555555, 0x00000001, 31, 1), 0xF5555555);
|
|
}
|
|
|
|
uint32_t run_Ext(uint32_t source, uint16_t pos, uint16_t size) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
|
|
__ li(v0, 0xFFFFFFFF);
|
|
__ li(t0, source);
|
|
__ Ext(v0, t0, pos, size);
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
auto f = GeneratedCode<F2>::FromCode(*code);
|
|
|
|
uint32_t res = reinterpret_cast<uint32_t>(f.Call(0, 0, 0, 0, 0));
|
|
|
|
return res;
|
|
}
|
|
|
|
TEST(Ext) {
|
|
CcTest::InitializeVM();
|
|
|
|
// Source values with negative sign.
|
|
// run_Ext(rs_value, pos, size), expected_result
|
|
CHECK_EQ(run_Ext(0x80000001, 0, 1), 0x00000001);
|
|
CHECK_EQ(run_Ext(0x80000001, 0, 32), 0x80000001);
|
|
CHECK_EQ(run_Ext(0x80000002, 1, 31), 0x40000001);
|
|
CHECK_EQ(run_Ext(0x80000100, 8, 24), 0x00800001);
|
|
CHECK_EQ(run_Ext(0x80010000, 16, 16), 0x00008001);
|
|
CHECK_EQ(run_Ext(0x81000000, 24, 8), 0x00000081);
|
|
CHECK_EQ(run_Ext(0x80000000, 31, 1), 0x00000001);
|
|
|
|
// Source values with positive sign.
|
|
CHECK_EQ(run_Ext(0x00000001, 0, 1), 0x00000001);
|
|
CHECK_EQ(run_Ext(0x40000001, 0, 32), 0x40000001);
|
|
CHECK_EQ(run_Ext(0x40000002, 1, 31), 0x20000001);
|
|
CHECK_EQ(run_Ext(0x40000100, 8, 24), 0x00400001);
|
|
CHECK_EQ(run_Ext(0x40010000, 16, 16), 0x00004001);
|
|
CHECK_EQ(run_Ext(0x41000000, 24, 8), 0x00000041);
|
|
CHECK_EQ(run_Ext(0x40000000, 31, 1), 0x00000000);
|
|
}
|
|
|
|
struct TestCaseMsaI5 {
|
|
uint64_t ws_lo;
|
|
uint64_t ws_hi;
|
|
uint32_t i5;
|
|
};
|
|
|
|
template <typename InstFunc, typename OperFunc>
|
|
void run_msa_i5(struct TestCaseMsaI5* input, bool i5_sign_ext,
|
|
InstFunc GenerateI5InstructionFunc,
|
|
OperFunc GenerateOperationFunc) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
CpuFeatureScope fscope(&assm, MIPS_SIMD);
|
|
msa_reg_t res;
|
|
int32_t i5 =
|
|
i5_sign_ext ? static_cast<int32_t>(input->i5 << 27) >> 27 : input->i5;
|
|
|
|
load_elements_of_vector(&assm, &(input->ws_lo), w0, t0, t1);
|
|
|
|
GenerateI5InstructionFunc(assm, i5);
|
|
|
|
store_elements_of_vector(&assm, w2, a0);
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
#ifdef OBJECT_PRINT
|
|
code->Print(std::cout);
|
|
#endif
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
|
|
(f.Call(&res, 0, 0, 0, 0));
|
|
|
|
CHECK_EQ(GenerateOperationFunc(input->ws_lo, input->i5), res.d[0]);
|
|
CHECK_EQ(GenerateOperationFunc(input->ws_hi, input->i5), res.d[1]);
|
|
}
|
|
|
|
TEST(MSA_addvi_subvi) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseMsaI5 tc[] = {
|
|
// ws_lo, ws_hi, i5
|
|
{0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0x0000001F},
|
|
{0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0x0000000F},
|
|
{0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0x00000005},
|
|
{0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0x00000010},
|
|
{0xFFAB807F807FFFCD, 0x7F23FF80FF567F80, 0x0000000F},
|
|
{0x80FFEFFF7F12807F, 0x807F80FF7FDEFF78, 0x00000010}};
|
|
|
|
#define ADDVI_DF(lanes, mask) \
|
|
uint64_t res = 0; \
|
|
for (int i = 0; i < lanes / 2; ++i) { \
|
|
int shift = (kMSARegSize / lanes) * i; \
|
|
res |= ((((ws >> shift) & mask) + i5) & mask) << shift; \
|
|
} \
|
|
return res
|
|
|
|
#define SUBVI_DF(lanes, mask) \
|
|
uint64_t res = 0; \
|
|
for (int i = 0; i < lanes / 2; ++i) { \
|
|
int shift = (kMSARegSize / lanes) * i; \
|
|
res |= ((((ws >> shift) & mask) - i5) & mask) << shift; \
|
|
} \
|
|
return res
|
|
|
|
for (size_t i = 0; i < sizeof(tc) / sizeof(TestCaseMsaI5); ++i) {
|
|
run_msa_i5(
|
|
&tc[i], false,
|
|
[](MacroAssembler& assm, int32_t i5) { __ addvi_b(w2, w0, i5); },
|
|
[](uint64_t ws, uint32_t i5) { ADDVI_DF(kMSALanesByte, UINT8_MAX); });
|
|
|
|
run_msa_i5(
|
|
&tc[i], false,
|
|
[](MacroAssembler& assm, int32_t i5) { __ addvi_h(w2, w0, i5); },
|
|
[](uint64_t ws, uint32_t i5) { ADDVI_DF(kMSALanesHalf, UINT16_MAX); });
|
|
|
|
run_msa_i5(
|
|
&tc[i], false,
|
|
[](MacroAssembler& assm, int32_t i5) { __ addvi_w(w2, w0, i5); },
|
|
[](uint64_t ws, uint32_t i5) { ADDVI_DF(kMSALanesWord, UINT32_MAX); });
|
|
|
|
run_msa_i5(
|
|
&tc[i], false,
|
|
[](MacroAssembler& assm, int32_t i5) { __ addvi_d(w2, w0, i5); },
|
|
[](uint64_t ws, uint32_t i5) { ADDVI_DF(kMSALanesDword, UINT64_MAX); });
|
|
|
|
run_msa_i5(
|
|
&tc[i], false,
|
|
[](MacroAssembler& assm, int32_t i5) { __ subvi_b(w2, w0, i5); },
|
|
[](uint64_t ws, uint32_t i5) { SUBVI_DF(kMSALanesByte, UINT8_MAX); });
|
|
|
|
run_msa_i5(
|
|
&tc[i], false,
|
|
[](MacroAssembler& assm, int32_t i5) { __ subvi_h(w2, w0, i5); },
|
|
[](uint64_t ws, uint32_t i5) { SUBVI_DF(kMSALanesHalf, UINT16_MAX); });
|
|
|
|
run_msa_i5(
|
|
&tc[i], false,
|
|
[](MacroAssembler& assm, int32_t i5) { __ subvi_w(w2, w0, i5); },
|
|
[](uint64_t ws, uint32_t i5) { SUBVI_DF(kMSALanesWord, UINT32_MAX); });
|
|
|
|
run_msa_i5(
|
|
&tc[i], false,
|
|
[](MacroAssembler& assm, int32_t i5) { __ subvi_d(w2, w0, i5); },
|
|
[](uint64_t ws, uint32_t i5) { SUBVI_DF(kMSALanesDword, UINT64_MAX); });
|
|
}
|
|
#undef ADDVI_DF
|
|
#undef SUBVI_DF
|
|
}
|
|
|
|
TEST(MSA_maxi_mini) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseMsaI5 tc[] = {
|
|
// ws_lo, ws_hi, i5
|
|
{0x7F80FF3480FF7F00, 0x8D7FFF80FF7F6780, 0x0000001F},
|
|
{0x7F80FF3480FF7F00, 0x8D7FFF80FF7F6780, 0x0000000F},
|
|
{0x7F80FF3480FF7F00, 0x8D7FFF80FF7F6780, 0x00000010},
|
|
{0x80007FFF91DAFFFF, 0x7FFF8000FFFF5678, 0x0000001F},
|
|
{0x80007FFF91DAFFFF, 0x7FFF8000FFFF5678, 0x0000000F},
|
|
{0x80007FFF91DAFFFF, 0x7FFF8000FFFF5678, 0x00000010},
|
|
{0x7FFFFFFF80000000, 0x12345678FFFFFFFF, 0x0000001F},
|
|
{0x7FFFFFFF80000000, 0x12345678FFFFFFFF, 0x0000000F},
|
|
{0x7FFFFFFF80000000, 0x12345678FFFFFFFF, 0x00000010},
|
|
{0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0x0000001F},
|
|
{0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0x0000000F},
|
|
{0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636, 0x00000010},
|
|
{0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0x00000015},
|
|
{0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0x00000009},
|
|
{0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636, 0x00000003}};
|
|
|
|
#define MAXI_MINI_S_DF(lanes, mask, func) \
|
|
[](uint64_t ws, uint32_t ui5) { \
|
|
uint64_t res = 0; \
|
|
int64_t i5 = ArithmeticShiftRight(static_cast<int64_t>(ui5) << 59, 59); \
|
|
int elem_size = kMSARegSize / lanes; \
|
|
for (int i = 0; i < lanes / 2; ++i) { \
|
|
int shift = elem_size * i; \
|
|
int64_t elem = \
|
|
static_cast<int64_t>(((ws >> shift) & mask) << (64 - elem_size)) >> \
|
|
(64 - elem_size); \
|
|
res |= static_cast<uint64_t>(func(elem, i5) & mask) << shift; \
|
|
} \
|
|
return res; \
|
|
}
|
|
|
|
#define MAXI_MINI_U_DF(lanes, mask, func) \
|
|
[](uint64_t ws, uint32_t ui5) { \
|
|
uint64_t res = 0; \
|
|
int elem_size = kMSARegSize / lanes; \
|
|
for (int i = 0; i < lanes / 2; ++i) { \
|
|
int shift = elem_size * i; \
|
|
uint64_t elem = (ws >> shift) & mask; \
|
|
res |= (func(elem, static_cast<uint64_t>(ui5)) & mask) << shift; \
|
|
} \
|
|
return res; \
|
|
}
|
|
|
|
for (size_t i = 0; i < sizeof(tc) / sizeof(TestCaseMsaI5); ++i) {
|
|
run_msa_i5(
|
|
&tc[i], true,
|
|
[](MacroAssembler& assm, int32_t i5) { __ maxi_s_b(w2, w0, i5); },
|
|
MAXI_MINI_S_DF(kMSALanesByte, UINT8_MAX, Max));
|
|
|
|
run_msa_i5(
|
|
&tc[i], true,
|
|
[](MacroAssembler& assm, int32_t i5) { __ maxi_s_h(w2, w0, i5); },
|
|
MAXI_MINI_S_DF(kMSALanesHalf, UINT16_MAX, Max));
|
|
|
|
run_msa_i5(
|
|
&tc[i], true,
|
|
[](MacroAssembler& assm, int32_t i5) { __ maxi_s_w(w2, w0, i5); },
|
|
MAXI_MINI_S_DF(kMSALanesWord, UINT32_MAX, Max));
|
|
|
|
run_msa_i5(
|
|
&tc[i], true,
|
|
[](MacroAssembler& assm, int32_t i5) { __ maxi_s_d(w2, w0, i5); },
|
|
MAXI_MINI_S_DF(kMSALanesDword, UINT64_MAX, Max));
|
|
|
|
run_msa_i5(
|
|
&tc[i], true,
|
|
[](MacroAssembler& assm, int32_t i5) { __ mini_s_b(w2, w0, i5); },
|
|
MAXI_MINI_S_DF(kMSALanesByte, UINT8_MAX, Min));
|
|
|
|
run_msa_i5(
|
|
&tc[i], true,
|
|
[](MacroAssembler& assm, int32_t i5) { __ mini_s_h(w2, w0, i5); },
|
|
MAXI_MINI_S_DF(kMSALanesHalf, UINT16_MAX, Min));
|
|
|
|
run_msa_i5(
|
|
&tc[i], true,
|
|
[](MacroAssembler& assm, int32_t i5) { __ mini_s_w(w2, w0, i5); },
|
|
MAXI_MINI_S_DF(kMSALanesWord, UINT32_MAX, Min));
|
|
|
|
run_msa_i5(
|
|
&tc[i], true,
|
|
[](MacroAssembler& assm, int32_t i5) { __ mini_s_d(w2, w0, i5); },
|
|
MAXI_MINI_S_DF(kMSALanesDword, UINT64_MAX, Min));
|
|
|
|
run_msa_i5(
|
|
&tc[i], false,
|
|
[](MacroAssembler& assm, int32_t i5) { __ maxi_u_b(w2, w0, i5); },
|
|
MAXI_MINI_U_DF(kMSALanesByte, UINT8_MAX, Max));
|
|
|
|
run_msa_i5(
|
|
&tc[i], false,
|
|
[](MacroAssembler& assm, int32_t i5) { __ maxi_u_h(w2, w0, i5); },
|
|
MAXI_MINI_U_DF(kMSALanesHalf, UINT16_MAX, Max));
|
|
|
|
run_msa_i5(
|
|
&tc[i], false,
|
|
[](MacroAssembler& assm, int32_t i5) { __ maxi_u_w(w2, w0, i5); },
|
|
MAXI_MINI_U_DF(kMSALanesWord, UINT32_MAX, Max));
|
|
|
|
run_msa_i5(
|
|
&tc[i], false,
|
|
[](MacroAssembler& assm, int32_t i5) { __ maxi_u_d(w2, w0, i5); },
|
|
MAXI_MINI_U_DF(kMSALanesDword, UINT64_MAX, Max));
|
|
|
|
run_msa_i5(
|
|
&tc[i], false,
|
|
[](MacroAssembler& assm, int32_t i5) { __ mini_u_b(w2, w0, i5); },
|
|
MAXI_MINI_U_DF(kMSALanesByte, UINT8_MAX, Min));
|
|
|
|
run_msa_i5(
|
|
&tc[i], false,
|
|
[](MacroAssembler& assm, int32_t i5) { __ mini_u_h(w2, w0, i5); },
|
|
MAXI_MINI_U_DF(kMSALanesHalf, UINT16_MAX, Min));
|
|
|
|
run_msa_i5(
|
|
&tc[i], false,
|
|
[](MacroAssembler& assm, int32_t i5) { __ mini_u_w(w2, w0, i5); },
|
|
MAXI_MINI_U_DF(kMSALanesWord, UINT32_MAX, Min));
|
|
|
|
run_msa_i5(
|
|
&tc[i], false,
|
|
[](MacroAssembler& assm, int32_t i5) { __ mini_u_d(w2, w0, i5); },
|
|
MAXI_MINI_U_DF(kMSALanesDword, UINT64_MAX, Min));
|
|
}
|
|
#undef MAXI_MINI_S_DF
|
|
#undef MAXI_MINI_U_DF
|
|
}
|
|
|
|
TEST(MSA_ceqi_clti_clei) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseMsaI5 tc[] = {
|
|
{0xFF69751BB9A7D9C3, 0xF7A594AEC8FF8A9C, 0x0000001F},
|
|
{0xE669FFFFB9A7D9C3, 0xF7A594AEFFFF8A9C, 0x0000001F},
|
|
{0xFFFFFFFFB9A7D9C3, 0xF7A594AEFFFFFFFF, 0x0000001F},
|
|
{0x2B0B5362C4E812DF, 0x3A0D80D68B3F0BC8, 0x0000000B},
|
|
{0x2B66000BC4E812DF, 0x3A0D000B8B3F8BC8, 0x0000000B},
|
|
{0x0000000BC4E812DF, 0x3A0D80D60000000B, 0x0000000B},
|
|
{0xF38062E13E38F8B0, 0x8041FFDEF2BFE636, 0x00000010},
|
|
{0xF35880003E38F8B0, 0x4F41FFDEF2BF8000, 0x00000010},
|
|
{0xF35862E180000000, 0x80000000F2BFE636, 0x00000010},
|
|
{0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0x00000015},
|
|
{0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0x00000009},
|
|
{0xF30062E13E38F800, 0x4F00FFDEF2BF0036, 0x00000000}};
|
|
|
|
#define CEQI_CLTI_CLEI_S_DF(lanes, mask, func) \
|
|
[](uint64_t ws, uint32_t ui5) { \
|
|
uint64_t res = 0; \
|
|
int elem_size = kMSARegSize / lanes; \
|
|
int64_t i5 = ArithmeticShiftRight(static_cast<int64_t>(ui5) << 59, 59); \
|
|
for (int i = 0; i < lanes / 2; ++i) { \
|
|
int shift = elem_size * i; \
|
|
int64_t elem = \
|
|
static_cast<int64_t>(((ws >> shift) & mask) << (64 - elem_size)) >> \
|
|
(64 - elem_size); \
|
|
res |= static_cast<uint64_t>((func)&mask) << shift; \
|
|
} \
|
|
return res; \
|
|
}
|
|
|
|
#define CEQI_CLTI_CLEI_U_DF(lanes, mask, func) \
|
|
[](uint64_t ws, uint64_t ui5) { \
|
|
uint64_t res = 0; \
|
|
int elem_size = kMSARegSize / lanes; \
|
|
for (int i = 0; i < lanes / 2; ++i) { \
|
|
int shift = elem_size * i; \
|
|
uint64_t elem = (ws >> shift) & mask; \
|
|
res |= ((func)&mask) << shift; \
|
|
} \
|
|
return res; \
|
|
}
|
|
|
|
for (size_t i = 0; i < sizeof(tc) / sizeof(TestCaseMsaI5); ++i) {
|
|
run_msa_i5(&tc[i], true,
|
|
[](MacroAssembler& assm, int32_t i5) { __ ceqi_b(w2, w0, i5); },
|
|
CEQI_CLTI_CLEI_S_DF(kMSALanesByte, UINT8_MAX,
|
|
!Compare(elem, i5) ? -1u : 0u));
|
|
|
|
run_msa_i5(&tc[i], true,
|
|
[](MacroAssembler& assm, int32_t i5) { __ ceqi_h(w2, w0, i5); },
|
|
CEQI_CLTI_CLEI_S_DF(kMSALanesHalf, UINT16_MAX,
|
|
!Compare(elem, i5) ? -1u : 0u));
|
|
|
|
run_msa_i5(&tc[i], true,
|
|
[](MacroAssembler& assm, int32_t i5) { __ ceqi_w(w2, w0, i5); },
|
|
CEQI_CLTI_CLEI_S_DF(kMSALanesWord, UINT32_MAX,
|
|
!Compare(elem, i5) ? -1u : 0u));
|
|
|
|
run_msa_i5(&tc[i], true,
|
|
[](MacroAssembler& assm, int32_t i5) { __ ceqi_d(w2, w0, i5); },
|
|
CEQI_CLTI_CLEI_S_DF(kMSALanesDword, UINT64_MAX,
|
|
!Compare(elem, i5) ? -1u : 0u));
|
|
|
|
run_msa_i5(
|
|
&tc[i], true,
|
|
[](MacroAssembler& assm, int32_t i5) { __ clti_s_b(w2, w0, i5); },
|
|
CEQI_CLTI_CLEI_S_DF(kMSALanesByte, UINT8_MAX,
|
|
(Compare(elem, i5) == -1) ? -1u : 0u));
|
|
|
|
run_msa_i5(
|
|
&tc[i], true,
|
|
[](MacroAssembler& assm, int32_t i5) { __ clti_s_h(w2, w0, i5); },
|
|
CEQI_CLTI_CLEI_S_DF(kMSALanesHalf, UINT16_MAX,
|
|
(Compare(elem, i5) == -1) ? -1u : 0u));
|
|
|
|
run_msa_i5(
|
|
&tc[i], true,
|
|
[](MacroAssembler& assm, int32_t i5) { __ clti_s_w(w2, w0, i5); },
|
|
CEQI_CLTI_CLEI_S_DF(kMSALanesWord, UINT32_MAX,
|
|
(Compare(elem, i5) == -1) ? -1u : 0u));
|
|
|
|
run_msa_i5(
|
|
&tc[i], true,
|
|
[](MacroAssembler& assm, int32_t i5) { __ clti_s_d(w2, w0, i5); },
|
|
CEQI_CLTI_CLEI_S_DF(kMSALanesDword, UINT64_MAX,
|
|
(Compare(elem, i5) == -1) ? -1ull : 0ull));
|
|
|
|
run_msa_i5(
|
|
&tc[i], true,
|
|
[](MacroAssembler& assm, int32_t i5) { __ clei_s_b(w2, w0, i5); },
|
|
CEQI_CLTI_CLEI_S_DF(kMSALanesByte, UINT8_MAX,
|
|
(Compare(elem, i5) != 1) ? -1u : 0u));
|
|
|
|
run_msa_i5(
|
|
&tc[i], true,
|
|
[](MacroAssembler& assm, int32_t i5) { __ clei_s_h(w2, w0, i5); },
|
|
CEQI_CLTI_CLEI_S_DF(kMSALanesHalf, UINT16_MAX,
|
|
(Compare(elem, i5) != 1) ? -1u : 0u));
|
|
|
|
run_msa_i5(
|
|
&tc[i], true,
|
|
[](MacroAssembler& assm, int32_t i5) { __ clei_s_w(w2, w0, i5); },
|
|
CEQI_CLTI_CLEI_S_DF(kMSALanesWord, UINT32_MAX,
|
|
(Compare(elem, i5) != 1) ? -1u : 0u));
|
|
|
|
run_msa_i5(
|
|
&tc[i], true,
|
|
[](MacroAssembler& assm, int32_t i5) { __ clei_s_d(w2, w0, i5); },
|
|
CEQI_CLTI_CLEI_S_DF(kMSALanesDword, UINT64_MAX,
|
|
(Compare(elem, i5) != 1) ? -1ull : 0ull));
|
|
|
|
run_msa_i5(
|
|
&tc[i], false,
|
|
[](MacroAssembler& assm, int32_t i5) { __ clti_u_b(w2, w0, i5); },
|
|
CEQI_CLTI_CLEI_U_DF(kMSALanesByte, UINT8_MAX,
|
|
(Compare(elem, ui5) == -1) ? -1ull : 0ull));
|
|
|
|
run_msa_i5(
|
|
&tc[i], false,
|
|
[](MacroAssembler& assm, int32_t i5) { __ clti_u_h(w2, w0, i5); },
|
|
CEQI_CLTI_CLEI_U_DF(kMSALanesHalf, UINT16_MAX,
|
|
(Compare(elem, ui5) == -1) ? -1ull : 0ull));
|
|
|
|
run_msa_i5(
|
|
&tc[i], false,
|
|
[](MacroAssembler& assm, int32_t i5) { __ clti_u_w(w2, w0, i5); },
|
|
CEQI_CLTI_CLEI_U_DF(kMSALanesWord, UINT32_MAX,
|
|
(Compare(elem, ui5) == -1) ? -1ull : 0ull));
|
|
|
|
run_msa_i5(
|
|
&tc[i], false,
|
|
[](MacroAssembler& assm, int32_t i5) { __ clti_u_d(w2, w0, i5); },
|
|
CEQI_CLTI_CLEI_U_DF(kMSALanesDword, UINT64_MAX,
|
|
(Compare(elem, ui5) == -1) ? -1ull : 0ull));
|
|
|
|
run_msa_i5(
|
|
&tc[i], false,
|
|
[](MacroAssembler& assm, int32_t i5) { __ clei_u_b(w2, w0, i5); },
|
|
CEQI_CLTI_CLEI_U_DF(kMSALanesByte, UINT8_MAX,
|
|
(Compare(elem, ui5) != 1) ? -1ull : 0ull));
|
|
|
|
run_msa_i5(
|
|
&tc[i], false,
|
|
[](MacroAssembler& assm, int32_t i5) { __ clei_u_h(w2, w0, i5); },
|
|
CEQI_CLTI_CLEI_U_DF(kMSALanesHalf, UINT16_MAX,
|
|
(Compare(elem, ui5) != 1) ? -1ull : 0ull));
|
|
|
|
run_msa_i5(
|
|
&tc[i], false,
|
|
[](MacroAssembler& assm, int32_t i5) { __ clei_u_w(w2, w0, i5); },
|
|
CEQI_CLTI_CLEI_U_DF(kMSALanesWord, UINT32_MAX,
|
|
(Compare(elem, ui5) != 1) ? -1ull : 0ull));
|
|
|
|
run_msa_i5(
|
|
&tc[i], false,
|
|
[](MacroAssembler& assm, int32_t i5) { __ clei_u_d(w2, w0, i5); },
|
|
CEQI_CLTI_CLEI_U_DF(kMSALanesDword, UINT64_MAX,
|
|
(Compare(elem, ui5) != 1) ? -1ull : 0ull));
|
|
}
|
|
#undef CEQI_CLTI_CLEI_S_DF
|
|
#undef CEQI_CLTI_CLEI_U_DF
|
|
}
|
|
|
|
struct TestCaseMsa2R {
|
|
uint64_t ws_lo;
|
|
uint64_t ws_hi;
|
|
uint64_t exp_res_lo;
|
|
uint64_t exp_res_hi;
|
|
};
|
|
|
|
template <typename Func>
|
|
void run_msa_2r(const struct TestCaseMsa2R* input,
|
|
Func Generate2RInstructionFunc) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
CpuFeatureScope fscope(&assm, MIPS_SIMD);
|
|
msa_reg_t res;
|
|
|
|
load_elements_of_vector(&assm, reinterpret_cast<const uint64_t*>(input), w0,
|
|
t0, t1);
|
|
Generate2RInstructionFunc(assm);
|
|
store_elements_of_vector(&assm, w2, a0);
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
#ifdef OBJECT_PRINT
|
|
code->Print(std::cout);
|
|
#endif
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
|
|
(f.Call(&res, 0, 0, 0, 0));
|
|
|
|
CHECK_EQ(input->exp_res_lo, res.d[0]);
|
|
CHECK_EQ(input->exp_res_hi, res.d[1]);
|
|
}
|
|
|
|
TEST(MSA_pcnt) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseMsa2R tc_b[] = {// ws_lo, ws_hi, exp_res_lo, exp_res_hi
|
|
{0x0000000000000000, 0x0000000000000000, 0, 0},
|
|
{0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF,
|
|
0x0808080808080808, 0x0808080808080808},
|
|
{0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C,
|
|
0x0204050405050504, 0x0704030503070304},
|
|
{0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8,
|
|
0x0404040303040207, 0x0403010504060403},
|
|
{0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636,
|
|
0x0603030405030503, 0x0502080605070504}};
|
|
|
|
struct TestCaseMsa2R tc_h[] = {// ws_lo, ws_hi, exp_res_lo, exp_res_hi
|
|
{0x0000000000000000, 0x0000000000000000, 0, 0},
|
|
{0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF,
|
|
0x0010001000100010, 0x0010001000100010},
|
|
{0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C,
|
|
0x00060009000A0009, 0x000B0008000A0007},
|
|
{0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8,
|
|
0x0008000700070009, 0x00070006000A0007},
|
|
{0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636,
|
|
0x0009000700080008, 0x0007000E000C0009}};
|
|
|
|
struct TestCaseMsa2R tc_w[] = {// ws_lo, ws_hi, exp_res_lo, exp_res_hi
|
|
{0x0000000000000000, 0x0000000000000000, 0, 0},
|
|
{0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF,
|
|
0x0000002000000020, 0x0000002000000020},
|
|
{0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C,
|
|
0x0000000F00000013, 0x0000001300000011},
|
|
{0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8,
|
|
0x0000000F00000010, 0x0000000D00000011},
|
|
{0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636,
|
|
0x0000001000000010, 0x0000001500000015}};
|
|
|
|
struct TestCaseMsa2R tc_d[] = {
|
|
// ws_lo, ws_hi, exp_res_lo, exp_res_hi
|
|
{0x0000000000000000, 0x0000000000000000, 0, 0},
|
|
{0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF, 0x40, 0x40},
|
|
{0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0x22, 0x24},
|
|
{0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0x1F, 0x1E},
|
|
{0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636, 0x20, 0x2A}};
|
|
|
|
for (size_t i = 0; i < sizeof(tc_b) / sizeof(TestCaseMsa2R); ++i) {
|
|
run_msa_2r(&tc_b[i], [](MacroAssembler& assm) { __ pcnt_b(w2, w0); });
|
|
run_msa_2r(&tc_h[i], [](MacroAssembler& assm) { __ pcnt_h(w2, w0); });
|
|
run_msa_2r(&tc_w[i], [](MacroAssembler& assm) { __ pcnt_w(w2, w0); });
|
|
run_msa_2r(&tc_d[i], [](MacroAssembler& assm) { __ pcnt_d(w2, w0); });
|
|
}
|
|
}
|
|
|
|
TEST(MSA_nlzc) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseMsa2R tc_b[] = {// ws_lo, ws_hi, exp_res_lo, exp_res_hi
|
|
{0x0000000000000000, 0x0000000000000000,
|
|
0x0808080808080808, 0x0808080808080808},
|
|
{0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF, 0, 0},
|
|
{0x1169350B07030100, 0x7F011402381F0A6C,
|
|
0x0301020405060708, 0x0107030602030401},
|
|
{0x010806003478121F, 0x03013016073F7B08,
|
|
0x0704050802010303, 0x0607020305020104},
|
|
{0x0168321100083803, 0x07113F03013F1676,
|
|
0x0701020308040206, 0x0503020607020301}};
|
|
|
|
struct TestCaseMsa2R tc_h[] = {// ws_lo, ws_hi, exp_res_lo, exp_res_hi
|
|
{0x0000000000000000, 0x0000000000000000,
|
|
0x0010001000100010, 0x0010001000100010},
|
|
{0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF, 0, 0},
|
|
{0x00010007000A003C, 0x37A5001E00010002,
|
|
0x000F000D000C000A, 0x0002000B000F000E},
|
|
{0x0026066200780EDF, 0x003D0003000F00C8,
|
|
0x000A000500090004, 0x000A000E000C0008},
|
|
{0x335807E100480030, 0x01410FDE12BF5636,
|
|
0x000200050009000A, 0x0007000400030001}};
|
|
|
|
struct TestCaseMsa2R tc_w[] = {// ws_lo, ws_hi, exp_res_lo, exp_res_hi
|
|
{0x0000000000000000, 0x0000000000000000,
|
|
0x0000002000000020, 0x0000002000000020},
|
|
{0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF, 0, 0},
|
|
{0x00000005000007C3, 0x000014AE00006A9C,
|
|
0x0000001D00000015, 0x0000001300000011},
|
|
{0x00009362000112DF, 0x000380D6003F8BC8,
|
|
0x000000100000000F, 0x0000000E0000000A},
|
|
{0x135862E17E38F8B0, 0x0061FFDE03BFE636,
|
|
0x0000000300000001, 0x0000000900000006}};
|
|
|
|
struct TestCaseMsa2R tc_d[] = {
|
|
// ws_lo, ws_hi, exp_res_lo, exp_res_hi
|
|
{0x0000000000000000, 0x0000000000000000, 0x40, 0x40},
|
|
{0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF, 0, 0},
|
|
{0x000000000000014E, 0x00000000000176DA, 0x37, 0x2F},
|
|
{0x00000062C4E812DF, 0x000065D68B3F8BC8, 0x19, 0x11},
|
|
{0x00000000E338F8B0, 0x0754534ACAB32654, 0x20, 0x5}};
|
|
|
|
for (size_t i = 0; i < sizeof(tc_b) / sizeof(TestCaseMsa2R); ++i) {
|
|
run_msa_2r(&tc_b[i], [](MacroAssembler& assm) { __ nlzc_b(w2, w0); });
|
|
run_msa_2r(&tc_h[i], [](MacroAssembler& assm) { __ nlzc_h(w2, w0); });
|
|
run_msa_2r(&tc_w[i], [](MacroAssembler& assm) { __ nlzc_w(w2, w0); });
|
|
run_msa_2r(&tc_d[i], [](MacroAssembler& assm) { __ nlzc_d(w2, w0); });
|
|
}
|
|
}
|
|
|
|
TEST(MSA_nloc) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseMsa2R tc_b[] = {// ws_lo, ws_hi, exp_res_lo, exp_res_hi
|
|
{0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF,
|
|
0x0808080808080808, 0x0808080808080808},
|
|
{0x0000000000000000, 0x0000000000000000, 0, 0},
|
|
{0xEE96CAF4F8FCFEFF, 0x80FEEBFDC7E0F593,
|
|
0x0301020405060708, 0x0107030602030401},
|
|
{0xFEF7F9FFCB87EDE0, 0xFCFECFE9F8C084F7,
|
|
0x0704050802010303, 0x0607020305020104},
|
|
{0xFE97CDEEFFF7C7FC, 0xF8EEC0FCFEC0E989,
|
|
0x0701020308040206, 0x0503020607020301}};
|
|
|
|
struct TestCaseMsa2R tc_h[] = {// ws_lo, ws_hi, exp_res_lo, exp_res_hi
|
|
{0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF,
|
|
0x0010001000100010, 0x0010001000100010},
|
|
{0x0000000000000000, 0x0000000000000000, 0, 0},
|
|
{0xFFFEFFF8FFF5FFC3, 0xC85AFFE1FFFEFFFD,
|
|
0x000F000D000C000A, 0x0002000B000F000E},
|
|
{0xFFD9F99DFF87F120, 0xFFC2FFFCFFF0FF37,
|
|
0x000A000500090004, 0x000A000E000C0008},
|
|
{0xCCA7F81EFFB7FFCF, 0xFEBEF021ED40A9C9,
|
|
0x000200050009000A, 0x0007000400030001}};
|
|
|
|
struct TestCaseMsa2R tc_w[] = {// ws_lo, ws_hi, exp_res_lo, exp_res_hi
|
|
{0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF,
|
|
0x0000002000000020, 0x0000002000000020},
|
|
{0x0000000000000000, 0x0000000000000000, 0, 0},
|
|
{0xFFFFFFFAFFFFF83C, 0xFFFFEB51FFFF9563,
|
|
0x0000001D00000015, 0x0000001300000011},
|
|
{0xFFFF6C9DFFFEED20, 0xFFFC7F29FFC07437,
|
|
0x000000100000000F, 0x0000000E0000000A},
|
|
{0xECA79D1E81C7074F, 0xFF9E0021FC4019C9,
|
|
0x0000000300000001, 0x0000000900000006}};
|
|
|
|
struct TestCaseMsa2R tc_d[] = {
|
|
// ws_lo, ws_hi, exp_res_lo, exp_res_hi
|
|
{0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF, 0x40, 0x40},
|
|
{0x0000000000000000, 0x0000000000000000, 0, 0},
|
|
{0xFFFFFFFFFFFFFEB1, 0xFFFFFFFFFFFE8925, 0x37, 0x2F},
|
|
{0xFFFFFF9D3B17ED20, 0xFFFF9A2974C07437, 0x19, 0x11},
|
|
{0xFFFFFFFF1CC7074F, 0xF8ABACB5354CD9AB, 0x20, 0x5}};
|
|
|
|
for (size_t i = 0; i < sizeof(tc_b) / sizeof(TestCaseMsa2R); ++i) {
|
|
run_msa_2r(&tc_b[i], [](MacroAssembler& assm) { __ nloc_b(w2, w0); });
|
|
run_msa_2r(&tc_h[i], [](MacroAssembler& assm) { __ nloc_h(w2, w0); });
|
|
run_msa_2r(&tc_w[i], [](MacroAssembler& assm) { __ nloc_w(w2, w0); });
|
|
run_msa_2r(&tc_d[i], [](MacroAssembler& assm) { __ nloc_d(w2, w0); });
|
|
}
|
|
}
|
|
|
|
struct TestCaseMsa2RF_F_U {
|
|
float ws1;
|
|
float ws2;
|
|
float ws3;
|
|
float ws4;
|
|
uint32_t exp_res_1;
|
|
uint32_t exp_res_2;
|
|
uint32_t exp_res_3;
|
|
uint32_t exp_res_4;
|
|
};
|
|
|
|
struct TestCaseMsa2RF_D_U {
|
|
double ws1;
|
|
double ws2;
|
|
uint64_t exp_res_1;
|
|
uint64_t exp_res_2;
|
|
};
|
|
|
|
TEST(MSA_fclass) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
#define BIT(n) (0x1 << n)
|
|
#define SNAN_BIT BIT(0)
|
|
#define QNAN_BIT BIT(1)
|
|
#define NEG_INFINITY_BIT BIT((2))
|
|
#define NEG_NORMAL_BIT BIT(3)
|
|
#define NEG_SUBNORMAL_BIT BIT(4)
|
|
#define NEG_ZERO_BIT BIT(5)
|
|
#define POS_INFINITY_BIT BIT(6)
|
|
#define POS_NORMAL_BIT BIT(7)
|
|
#define POS_SUBNORMAL_BIT BIT(8)
|
|
#define POS_ZERO_BIT BIT(9)
|
|
|
|
const float inf_float = std::numeric_limits<float>::infinity();
|
|
const double inf_double = std::numeric_limits<double>::infinity();
|
|
|
|
const struct TestCaseMsa2RF_F_U tc_s[] = {
|
|
{1.f, -0.00001, 208e10f, -34.8e-30f, POS_NORMAL_BIT, NEG_NORMAL_BIT,
|
|
POS_NORMAL_BIT, NEG_NORMAL_BIT},
|
|
{inf_float, -inf_float, 0, -0.f, POS_INFINITY_BIT, NEG_INFINITY_BIT,
|
|
POS_ZERO_BIT, NEG_ZERO_BIT},
|
|
{3.036e-40f, -6.392e-43f, 1.41e-45f, -1.17e-38f, POS_SUBNORMAL_BIT,
|
|
NEG_SUBNORMAL_BIT, POS_SUBNORMAL_BIT, NEG_SUBNORMAL_BIT}};
|
|
|
|
const struct TestCaseMsa2RF_D_U tc_d[] = {
|
|
{1., -0.00000001, POS_NORMAL_BIT, NEG_NORMAL_BIT},
|
|
{208e10, -34.8e-300, POS_NORMAL_BIT, NEG_NORMAL_BIT},
|
|
{inf_double, -inf_double, POS_INFINITY_BIT, NEG_INFINITY_BIT},
|
|
{0, -0., POS_ZERO_BIT, NEG_ZERO_BIT},
|
|
{1.036e-308, -6.392e-309, POS_SUBNORMAL_BIT, NEG_SUBNORMAL_BIT},
|
|
{1.41e-323, -3.17e208, POS_SUBNORMAL_BIT, NEG_NORMAL_BIT}};
|
|
|
|
for (size_t i = 0; i < sizeof(tc_s) / sizeof(TestCaseMsa2RF_F_U); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_s[i]),
|
|
[](MacroAssembler& assm) { __ fclass_w(w2, w0); });
|
|
}
|
|
for (size_t i = 0; i < sizeof(tc_d) / sizeof(TestCaseMsa2RF_D_U); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]),
|
|
[](MacroAssembler& assm) { __ fclass_d(w2, w0); });
|
|
}
|
|
|
|
#undef BIT
|
|
#undef SNAN_BIT
|
|
#undef QNAN_BIT
|
|
#undef NEG_INFINITY_BIT
|
|
#undef NEG_NORMAL_BIT
|
|
#undef NEG_SUBNORMAL_BIT
|
|
#undef NEG_ZERO_BIT
|
|
#undef POS_INFINITY_BIT
|
|
#undef POS_NORMAL_BIT
|
|
#undef POS_SUBNORMAL_BIT
|
|
#undef POS_ZERO_BIT
|
|
}
|
|
|
|
struct TestCaseMsa2RF_F_I {
|
|
float ws1;
|
|
float ws2;
|
|
float ws3;
|
|
float ws4;
|
|
int32_t exp_res_1;
|
|
int32_t exp_res_2;
|
|
int32_t exp_res_3;
|
|
int32_t exp_res_4;
|
|
};
|
|
|
|
struct TestCaseMsa2RF_D_I {
|
|
double ws1;
|
|
double ws2;
|
|
int64_t exp_res_1;
|
|
int64_t exp_res_2;
|
|
};
|
|
|
|
TEST(MSA_ftrunc_s) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
const float inf_float = std::numeric_limits<float>::infinity();
|
|
const float qNaN_float = std::numeric_limits<float>::quiet_NaN();
|
|
const double inf_double = std::numeric_limits<double>::infinity();
|
|
const double qNaN_double = std::numeric_limits<double>::quiet_NaN();
|
|
const int32_t max_int32 = std::numeric_limits<int32_t>::max();
|
|
const int32_t min_int32 = std::numeric_limits<int32_t>::min();
|
|
const int64_t max_int64 = std::numeric_limits<int64_t>::max();
|
|
const int64_t min_int64 = std::numeric_limits<int64_t>::min();
|
|
|
|
const struct TestCaseMsa2RF_F_I tc_s[] = {
|
|
{inf_float, 2.345f, -324.9235f, 30004.51f, max_int32, 2, -324, 30004},
|
|
{-inf_float, -0.983f, 0.0832f, static_cast<float>(max_int32) * 3.f,
|
|
min_int32, 0, 0, max_int32},
|
|
{-23.125f, qNaN_float, 2 * static_cast<float>(min_int32), -0.f, -23, 0,
|
|
min_int32, 0}};
|
|
|
|
const struct TestCaseMsa2RF_D_I tc_d[] = {
|
|
{inf_double, 2.345, max_int64, 2},
|
|
{-324.9235, 246569139.51, -324, 246569139},
|
|
{-inf_double, -0.983, min_int64, 0},
|
|
{0.0832, 6 * static_cast<double>(max_int64), 0, max_int64},
|
|
{-21453889872.94, qNaN_double, -21453889872, 0},
|
|
{2 * static_cast<double>(min_int64), -0., min_int64, 0}};
|
|
|
|
for (size_t i = 0; i < sizeof(tc_s) / sizeof(TestCaseMsa2RF_F_I); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_s[i]),
|
|
[](MacroAssembler& assm) { __ ftrunc_s_w(w2, w0); });
|
|
}
|
|
for (size_t i = 0; i < sizeof(tc_d) / sizeof(TestCaseMsa2RF_D_I); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]),
|
|
[](MacroAssembler& assm) { __ ftrunc_s_d(w2, w0); });
|
|
}
|
|
}
|
|
|
|
TEST(MSA_ftrunc_u) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
const float inf_float = std::numeric_limits<float>::infinity();
|
|
const float qNaN_float = std::numeric_limits<float>::quiet_NaN();
|
|
const double inf_double = std::numeric_limits<double>::infinity();
|
|
const double qNaN_double = std::numeric_limits<double>::quiet_NaN();
|
|
const uint32_t max_uint32 = std::numeric_limits<uint32_t>::max();
|
|
const uint64_t max_uint64 = std::numeric_limits<uint64_t>::max();
|
|
|
|
const struct TestCaseMsa2RF_F_U tc_s[] = {
|
|
{inf_float, 2.345f, -324.9235f, 30004.51f, max_uint32, 2, 0, 30004},
|
|
{-inf_float, 0.983f, 0.0832f, static_cast<float>(max_uint32) * 3., 0, 0,
|
|
0, max_uint32},
|
|
{23.125f, qNaN_float, -0.982, -0.f, 23, 0, 0, 0}};
|
|
|
|
const struct TestCaseMsa2RF_D_U tc_d[] = {
|
|
{inf_double, 2.345, max_uint64, 2},
|
|
{-324.9235, 246569139.51, 0, 246569139},
|
|
{-inf_double, -0.983, 0, 0},
|
|
{0.0832, 6 * static_cast<double>(max_uint64), 0, max_uint64},
|
|
{21453889872.94, qNaN_double, 21453889872, 0},
|
|
{0.9889, -0., 0, 0}};
|
|
|
|
for (size_t i = 0; i < sizeof(tc_s) / sizeof(TestCaseMsa2RF_F_U); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_s[i]),
|
|
[](MacroAssembler& assm) { __ ftrunc_u_w(w2, w0); });
|
|
}
|
|
for (size_t i = 0; i < sizeof(tc_d) / sizeof(TestCaseMsa2RF_D_U); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]),
|
|
[](MacroAssembler& assm) { __ ftrunc_u_d(w2, w0); });
|
|
}
|
|
}
|
|
|
|
struct TestCaseMsa2RF_F_F {
|
|
float ws1;
|
|
float ws2;
|
|
float ws3;
|
|
float ws4;
|
|
float exp_res_1;
|
|
float exp_res_2;
|
|
float exp_res_3;
|
|
float exp_res_4;
|
|
};
|
|
|
|
struct TestCaseMsa2RF_D_D {
|
|
double ws1;
|
|
double ws2;
|
|
double exp_res_1;
|
|
double exp_res_2;
|
|
};
|
|
|
|
TEST(MSA_fsqrt) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
const float inf_float = std::numeric_limits<float>::infinity();
|
|
const double inf_double = std::numeric_limits<double>::infinity();
|
|
|
|
const struct TestCaseMsa2RF_F_F tc_s[] = {
|
|
{81.f, 576.f, inf_float, -0.f, 9.f, 24.f, inf_float, -0.f}};
|
|
|
|
const struct TestCaseMsa2RF_D_D tc_d[] = {{81., inf_double, 9., inf_double},
|
|
{331776., -0., 576, -0.}};
|
|
|
|
for (size_t i = 0; i < sizeof(tc_s) / sizeof(TestCaseMsa2RF_F_F); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_s[i]),
|
|
[](MacroAssembler& assm) { __ fsqrt_w(w2, w0); });
|
|
}
|
|
for (size_t i = 0; i < sizeof(tc_d) / sizeof(TestCaseMsa2RF_D_D); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]),
|
|
[](MacroAssembler& assm) { __ fsqrt_d(w2, w0); });
|
|
}
|
|
}
|
|
|
|
TEST(MSA_frsqrt) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
const float inf_float = std::numeric_limits<float>::infinity();
|
|
const double inf_double = std::numeric_limits<double>::infinity();
|
|
|
|
const struct TestCaseMsa2RF_F_F tc_s[] = {
|
|
{81.f, 576.f, inf_float, -0.f, 1.f / 9.f, 1.f / 24.f, 0.f, -inf_float},
|
|
{0.f, 1.f / 576.f, 1.f / 81.f, 1.f / 4.f, inf_float, 24.f, 9.f, 2.f}};
|
|
|
|
const struct TestCaseMsa2RF_D_D tc_d[] = {
|
|
{81., inf_double, 1. / 9., 0.},
|
|
{331776., -0., 1. / 576., -inf_double},
|
|
{0., 1. / 81, inf_double, 9.}};
|
|
|
|
for (size_t i = 0; i < sizeof(tc_s) / sizeof(TestCaseMsa2RF_F_F); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_s[i]),
|
|
[](MacroAssembler& assm) { __ frsqrt_w(w2, w0); });
|
|
}
|
|
for (size_t i = 0; i < sizeof(tc_d) / sizeof(TestCaseMsa2RF_D_D); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]),
|
|
[](MacroAssembler& assm) { __ frsqrt_d(w2, w0); });
|
|
}
|
|
}
|
|
|
|
TEST(MSA_frcp) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
const float inf_float = std::numeric_limits<float>::infinity();
|
|
const double inf_double = std::numeric_limits<double>::infinity();
|
|
|
|
const struct TestCaseMsa2RF_F_F tc_s[] = {
|
|
{12.f, 576.f, inf_float, -0.f, 1.f / 12.f, 1.f / 576.f, 0.f, -inf_float},
|
|
{0.f, 1.f / 576.f, -inf_float, 1.f / 400.f, inf_float, 576.f, -0.f,
|
|
400.f}};
|
|
|
|
const struct TestCaseMsa2RF_D_D tc_d[] = {
|
|
{81., inf_double, 1. / 81., 0.},
|
|
{331777., -0., 1. / 331777., -inf_double},
|
|
{0., 1. / 80, inf_double, 80.},
|
|
{1. / 40000., -inf_double, 40000., -0.}};
|
|
|
|
for (size_t i = 0; i < sizeof(tc_s) / sizeof(TestCaseMsa2RF_F_F); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_s[i]),
|
|
[](MacroAssembler& assm) { __ frcp_w(w2, w0); });
|
|
}
|
|
for (size_t i = 0; i < sizeof(tc_d) / sizeof(TestCaseMsa2RF_D_D); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]),
|
|
[](MacroAssembler& assm) { __ frcp_d(w2, w0); });
|
|
}
|
|
}
|
|
|
|
void test_frint_s(size_t data_size, TestCaseMsa2RF_F_F tc_d[],
|
|
int rounding_mode) {
|
|
for (size_t i = 0; i < data_size / sizeof(TestCaseMsa2RF_F_F); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]),
|
|
[&rounding_mode](MacroAssembler& assm) {
|
|
MSAControlRegister msareg = {kMSACSRRegister};
|
|
__ li(t0, static_cast<uint32_t>(rounding_mode));
|
|
__ cfcmsa(t1, msareg);
|
|
__ ctcmsa(msareg, t0);
|
|
__ frint_w(w2, w0);
|
|
__ ctcmsa(msareg, t1);
|
|
});
|
|
}
|
|
}
|
|
|
|
void test_frint_d(size_t data_size, TestCaseMsa2RF_D_D tc_d[],
|
|
int rounding_mode) {
|
|
for (size_t i = 0; i < data_size / sizeof(TestCaseMsa2RF_D_D); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]),
|
|
[&rounding_mode](MacroAssembler& assm) {
|
|
MSAControlRegister msareg = {kMSACSRRegister};
|
|
__ li(t0, static_cast<uint32_t>(rounding_mode));
|
|
__ cfcmsa(t1, msareg);
|
|
__ ctcmsa(msareg, t0);
|
|
__ frint_d(w2, w0);
|
|
__ ctcmsa(msareg, t1);
|
|
});
|
|
}
|
|
}
|
|
|
|
TEST(MSA_frint) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseMsa2RF_F_F tc_s1[] = {
|
|
{0.f, 4.51f, 1.49f, -12.51f, 0.f, 5.f, 1.f, -13.f},
|
|
{-1.32f, -23.38f, 2.8f, -32.5f, -1.f, -23.f, 3.f, -32.f}};
|
|
|
|
struct TestCaseMsa2RF_D_D tc_d1[] = {{0., 4.51, 0., 5.},
|
|
{1.49, -12.51, 1., -13.},
|
|
{-1.32, -23.38, -1., -23.},
|
|
{2.8, -32.6, 3., -33.}};
|
|
|
|
test_frint_s(sizeof(tc_s1), tc_s1, kRoundToNearest);
|
|
test_frint_d(sizeof(tc_d1), tc_d1, kRoundToNearest);
|
|
|
|
struct TestCaseMsa2RF_F_F tc_s2[] = {
|
|
{0.f, 4.5f, 1.49f, -12.51f, 0.f, 4.f, 1.f, -12.f},
|
|
{-1.f, -23.38f, 2.8f, -32.6f, -1.f, -23.f, 2.f, -32.f}};
|
|
|
|
struct TestCaseMsa2RF_D_D tc_d2[] = {{0., 4.5, 0., 4.},
|
|
{1.49, -12.51, 1., -12.},
|
|
{-1., -23.38, -1., -23.},
|
|
{2.8, -32.6, 2., -32.}};
|
|
|
|
test_frint_s(sizeof(tc_s2), tc_s2, kRoundToZero);
|
|
test_frint_d(sizeof(tc_d2), tc_d2, kRoundToZero);
|
|
|
|
struct TestCaseMsa2RF_F_F tc_s3[] = {
|
|
{0.f, 4.5f, 1.49f, -12.51f, 0.f, 5.f, 2.f, -12.f},
|
|
{-1.f, -23.38f, 2.8f, -32.6f, -1.f, -23.f, 3.f, -32.f}};
|
|
|
|
struct TestCaseMsa2RF_D_D tc_d3[] = {{0., 4.5, 0., 5.},
|
|
{1.49, -12.51, 2., -12.},
|
|
{-1., -23.38, -1., -23.},
|
|
{2.8, -32.6, 3., -32.}};
|
|
|
|
test_frint_s(sizeof(tc_s3), tc_s3, kRoundToPlusInf);
|
|
test_frint_d(sizeof(tc_d3), tc_d3, kRoundToPlusInf);
|
|
|
|
struct TestCaseMsa2RF_F_F tc_s4[] = {
|
|
{0.f, 4.5f, 1.49f, -12.51f, 0.f, 4.f, 1.f, -13.f},
|
|
{-1.f, -23.38f, 2.8f, -32.6f, -1.f, -24.f, 2.f, -33.f}};
|
|
|
|
struct TestCaseMsa2RF_D_D tc_d4[] = {{0., 4.5, 0., 4.},
|
|
{1.49, -12.51, 1., -13.},
|
|
{-1., -23.38, -1., -24.},
|
|
{2.8, -32.6, 2., -33.}};
|
|
|
|
test_frint_s(sizeof(tc_s4), tc_s4, kRoundToMinusInf);
|
|
test_frint_d(sizeof(tc_d4), tc_d4, kRoundToMinusInf);
|
|
}
|
|
|
|
TEST(MSA_flog2) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
const float inf_float = std::numeric_limits<float>::infinity();
|
|
const double inf_double = std::numeric_limits<double>::infinity();
|
|
|
|
struct TestCaseMsa2RF_F_F tc_s[] = {
|
|
{std::ldexp(0.58f, -48), std::ldexp(0.5f, 110), std::ldexp(1.11f, -130),
|
|
inf_float, -49.f, 109.f, -130.f, inf_float},
|
|
{0.f, -0.f, std::ldexp(0.89f, -12), std::ldexp(0.32f, 126), -inf_float,
|
|
-inf_float, -13.f, 124.f}};
|
|
|
|
struct TestCaseMsa2RF_D_D tc_d[] = {
|
|
{std::ldexp(0.58, -48), std::ldexp(0.5, 110), -49., 109.},
|
|
{std::ldexp(1.11, -1050), inf_double, -1050., inf_double},
|
|
{0., -0., -inf_double, -inf_double},
|
|
{std::ldexp(0.32, 1021), std::ldexp(1.23, -123), 1019., -123.}};
|
|
|
|
for (size_t i = 0; i < sizeof(tc_s) / sizeof(TestCaseMsa2RF_F_F); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_s[i]),
|
|
[](MacroAssembler& assm) { __ flog2_w(w2, w0); });
|
|
}
|
|
|
|
for (size_t i = 0; i < sizeof(tc_d) / sizeof(TestCaseMsa2RF_D_D); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]),
|
|
[](MacroAssembler& assm) { __ flog2_d(w2, w0); });
|
|
}
|
|
}
|
|
|
|
void test_ftint_s_s(size_t data_size, TestCaseMsa2RF_F_I tc_d[],
|
|
int rounding_mode) {
|
|
for (size_t i = 0; i < data_size / sizeof(TestCaseMsa2RF_F_I); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]),
|
|
[&rounding_mode](MacroAssembler& assm) {
|
|
MSAControlRegister msareg = {kMSACSRRegister};
|
|
__ li(t0, static_cast<uint32_t>(rounding_mode));
|
|
__ cfcmsa(t1, msareg);
|
|
__ ctcmsa(msareg, t0);
|
|
__ ftint_s_w(w2, w0);
|
|
__ ctcmsa(msareg, t1);
|
|
});
|
|
}
|
|
}
|
|
|
|
void test_ftint_s_d(size_t data_size, TestCaseMsa2RF_D_I tc_d[],
|
|
int rounding_mode) {
|
|
for (size_t i = 0; i < data_size / sizeof(TestCaseMsa2RF_D_I); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]),
|
|
[&rounding_mode](MacroAssembler& assm) {
|
|
MSAControlRegister msareg = {kMSACSRRegister};
|
|
__ li(t0, static_cast<uint32_t>(rounding_mode));
|
|
__ cfcmsa(t1, msareg);
|
|
__ ctcmsa(msareg, t0);
|
|
__ ftint_s_d(w2, w0);
|
|
__ ctcmsa(msareg, t1);
|
|
});
|
|
}
|
|
}
|
|
|
|
TEST(MSA_ftint_s) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
const float inf_float = std::numeric_limits<float>::infinity();
|
|
const double inf_double = std::numeric_limits<double>::infinity();
|
|
const int32_t int32_max = std::numeric_limits<int32_t>::max();
|
|
const int32_t int32_min = std::numeric_limits<int32_t>::min();
|
|
const int64_t int64_max = std::numeric_limits<int64_t>::max();
|
|
const int64_t int64_min = std::numeric_limits<int64_t>::min();
|
|
|
|
struct TestCaseMsa2RF_F_I tc_s1[] = {
|
|
{0.f, 4.51f, 1.49f, -12.51f, 0, 5, 1, -13},
|
|
{-0.32f, -23.38f, 2.8f, -32.6f, 0, -23, 3, -33},
|
|
{inf_float, -inf_float, 3.f * int32_min, 4.f * int32_max, int32_max,
|
|
int32_min, int32_min, int32_max}};
|
|
|
|
struct TestCaseMsa2RF_D_I tc_d1[] = {
|
|
{0., 4.51, 0, 5},
|
|
{1.49, -12.51, 1, -13},
|
|
{-0.32, -23.38, 0, -23},
|
|
{2.8, -32.6, 3, -33},
|
|
{inf_double, -inf_double, int64_max, int64_min},
|
|
{33.23 * int64_min, 4000. * int64_max, int64_min, int64_max}};
|
|
|
|
test_ftint_s_s(sizeof(tc_s1), tc_s1, kRoundToNearest);
|
|
test_ftint_s_d(sizeof(tc_d1), tc_d1, kRoundToNearest);
|
|
|
|
struct TestCaseMsa2RF_F_I tc_s2[] = {
|
|
{0.f, 4.5f, 1.49f, -12.51f, 0, 4, 1, -12},
|
|
{-0.f, -23.38f, 2.8f, -32.6f, -0, -23, 2, -32},
|
|
{inf_float, -inf_float, 3.f * int32_min, 4.f * int32_max, int32_max,
|
|
int32_min, int32_min, int32_max}};
|
|
|
|
struct TestCaseMsa2RF_D_I tc_d2[] = {
|
|
{0., 4.5, 0, 4},
|
|
{1.49, -12.51, 1, -12},
|
|
{-0., -23.38, -0, -23},
|
|
{2.8, -32.6, 2, -32},
|
|
{inf_double, -inf_double, int64_max, int64_min},
|
|
{33.23 * int64_min, 4000. * int64_max, int64_min, int64_max}};
|
|
|
|
test_ftint_s_s(sizeof(tc_s2), tc_s2, kRoundToZero);
|
|
test_ftint_s_d(sizeof(tc_d2), tc_d2, kRoundToZero);
|
|
|
|
struct TestCaseMsa2RF_F_I tc_s3[] = {
|
|
{0.f, 4.5f, 1.49f, -12.51f, 0, 5, 2, -12},
|
|
{-0.f, -23.38f, 2.8f, -32.6f, -0, -23, 3, -32},
|
|
{inf_float, -inf_float, 3.f * int32_min, 4.f * int32_max, int32_max,
|
|
int32_min, int32_min, int32_max}};
|
|
|
|
struct TestCaseMsa2RF_D_I tc_d3[] = {
|
|
{0., 4.5, 0, 5},
|
|
{1.49, -12.51, 2, -12},
|
|
{-0., -23.38, -0, -23},
|
|
{2.8, -32.6, 3, -32},
|
|
{inf_double, -inf_double, int64_max, int64_min},
|
|
{33.23 * int64_min, 4000. * int64_max, int64_min, int64_max}};
|
|
|
|
test_ftint_s_s(sizeof(tc_s3), tc_s3, kRoundToPlusInf);
|
|
test_ftint_s_d(sizeof(tc_d3), tc_d3, kRoundToPlusInf);
|
|
|
|
struct TestCaseMsa2RF_F_I tc_s4[] = {
|
|
{0.f, 4.5f, 1.49f, -12.51f, 0, 4, 1, -13},
|
|
{-0.f, -23.38f, 2.8f, -32.6f, -0, -24, 2, -33},
|
|
{inf_float, -inf_float, 3.f * int32_min, 4.f * int32_max, int32_max,
|
|
int32_min, int32_min, int32_max}};
|
|
|
|
struct TestCaseMsa2RF_D_I tc_d4[] = {
|
|
{0., 4.5, 0, 4},
|
|
{1.49, -12.51, 1, -13},
|
|
{-0., -23.38, -0, -24},
|
|
{2.8, -32.6, 2, -33},
|
|
{inf_double, -inf_double, int64_max, int64_min},
|
|
{33.23 * int64_min, 4000. * int64_max, int64_min, int64_max}};
|
|
|
|
test_ftint_s_s(sizeof(tc_s4), tc_s4, kRoundToMinusInf);
|
|
test_ftint_s_d(sizeof(tc_d4), tc_d4, kRoundToMinusInf);
|
|
}
|
|
|
|
void test_ftint_u_s(size_t data_size, TestCaseMsa2RF_F_U tc_d[],
|
|
int rounding_mode) {
|
|
for (size_t i = 0; i < data_size / sizeof(TestCaseMsa2RF_F_U); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]),
|
|
[&rounding_mode](MacroAssembler& assm) {
|
|
MSAControlRegister msareg = {kMSACSRRegister};
|
|
__ li(t0, static_cast<uint32_t>(rounding_mode));
|
|
__ cfcmsa(t1, msareg);
|
|
__ ctcmsa(msareg, t0);
|
|
__ ftint_u_w(w2, w0);
|
|
__ ctcmsa(msareg, t1);
|
|
});
|
|
}
|
|
}
|
|
|
|
void test_ftint_u_d(size_t data_size, TestCaseMsa2RF_D_U tc_d[],
|
|
int rounding_mode) {
|
|
for (size_t i = 0; i < data_size / sizeof(TestCaseMsa2RF_D_U); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]),
|
|
[&rounding_mode](MacroAssembler& assm) {
|
|
MSAControlRegister msareg = {kMSACSRRegister};
|
|
__ li(t0, static_cast<uint32_t>(rounding_mode));
|
|
__ cfcmsa(t1, msareg);
|
|
__ ctcmsa(msareg, t0);
|
|
__ ftint_u_d(w2, w0);
|
|
__ ctcmsa(msareg, t1);
|
|
});
|
|
}
|
|
}
|
|
|
|
TEST(MSA_ftint_u) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
const float inf_float = std::numeric_limits<float>::infinity();
|
|
const double inf_double = std::numeric_limits<double>::infinity();
|
|
const uint32_t uint32_max = std::numeric_limits<uint32_t>::max();
|
|
const uint64_t uint64_max = std::numeric_limits<uint64_t>::max();
|
|
|
|
struct TestCaseMsa2RF_F_U tc_s1[] = {
|
|
{0.f, 4.51f, 1.49f, -12.51f, 0, 5, 1, 0},
|
|
{-0.32f, 23.38f, 2.8f, 32.6f, 0, 23, 3, 33},
|
|
{inf_float, -inf_float, 0, 4.f * uint32_max, uint32_max, 0, 0,
|
|
uint32_max}};
|
|
|
|
struct TestCaseMsa2RF_D_U tc_d1[] = {
|
|
{0., 4.51, 0, 5},
|
|
{1.49, -12.51, 1, 0},
|
|
{-0.32, 23.38, 0, 23},
|
|
{2.8, 32.6, 3, 33},
|
|
{inf_double, -inf_double, uint64_max, 0},
|
|
{-0., 4000. * uint64_max, 0, uint64_max}};
|
|
|
|
test_ftint_u_s(sizeof(tc_s1), tc_s1, kRoundToNearest);
|
|
test_ftint_u_d(sizeof(tc_d1), tc_d1, kRoundToNearest);
|
|
|
|
struct TestCaseMsa2RF_F_U tc_s2[] = {
|
|
{0.f, 4.5f, 1.49f, -12.51f, 0, 4, 1, 0},
|
|
{-0.f, 23.38f, 2.8f, 32.6f, 0, 23, 2, 32},
|
|
{inf_float, -inf_float, 0., 4.f * uint32_max, uint32_max, 0, 0,
|
|
uint32_max}};
|
|
|
|
struct TestCaseMsa2RF_D_U tc_d2[] = {
|
|
{0., 4.5, 0, 4},
|
|
{1.49, -12.51, 1, 0},
|
|
{-0., 23.38, 0, 23},
|
|
{2.8, 32.6, 2, 32},
|
|
{inf_double, -inf_double, uint64_max, 0},
|
|
{-0.2345, 4000. * uint64_max, 0, uint64_max}};
|
|
|
|
test_ftint_u_s(sizeof(tc_s2), tc_s2, kRoundToZero);
|
|
test_ftint_u_d(sizeof(tc_d2), tc_d2, kRoundToZero);
|
|
|
|
struct TestCaseMsa2RF_F_U tc_s3[] = {
|
|
{0.f, 4.5f, 1.49f, -12.51f, 0, 5, 2, 0},
|
|
{-0.f, 23.38f, 2.8f, 32.6f, 0, 24, 3, 33},
|
|
{inf_float, -inf_float, 0, 4.f * uint32_max, uint32_max, 0, 0,
|
|
uint32_max}};
|
|
|
|
struct TestCaseMsa2RF_D_U tc_d3[] = {
|
|
{0., 4.5, 0, 5},
|
|
{1.49, -12.51, 2, 0},
|
|
{-0., 23.38, -0, 24},
|
|
{2.8, 32.6, 3, 33},
|
|
{inf_double, -inf_double, uint64_max, 0},
|
|
{-0.5252, 4000. * uint64_max, 0, uint64_max}};
|
|
|
|
test_ftint_u_s(sizeof(tc_s3), tc_s3, kRoundToPlusInf);
|
|
test_ftint_u_d(sizeof(tc_d3), tc_d3, kRoundToPlusInf);
|
|
|
|
struct TestCaseMsa2RF_F_U tc_s4[] = {
|
|
{0.f, 4.5f, 1.49f, -12.51f, 0, 4, 1, 0},
|
|
{-0.f, 23.38f, 2.8f, 32.6f, 0, 23, 2, 32},
|
|
{inf_float, -inf_float, 0, 4.f * uint32_max, uint32_max, 0, 0,
|
|
uint32_max}};
|
|
|
|
struct TestCaseMsa2RF_D_U tc_d4[] = {
|
|
{0., 4.5, 0, 4},
|
|
{1.49, -12.51, 1, 0},
|
|
{-0., 23.38, -0, 23},
|
|
{2.8, 32.6, 2, 32},
|
|
{inf_double, -inf_double, uint64_max, 0},
|
|
{-0.098797, 4000. * uint64_max, 0, uint64_max}};
|
|
|
|
test_ftint_u_s(sizeof(tc_s4), tc_s4, kRoundToMinusInf);
|
|
test_ftint_u_d(sizeof(tc_d4), tc_d4, kRoundToMinusInf);
|
|
}
|
|
|
|
struct TestCaseMsa2RF_U_F {
|
|
uint32_t ws1;
|
|
uint32_t ws2;
|
|
uint32_t ws3;
|
|
uint32_t ws4;
|
|
float exp_res_1;
|
|
float exp_res_2;
|
|
float exp_res_3;
|
|
float exp_res_4;
|
|
};
|
|
|
|
struct TestCaseMsa2RF_U_D {
|
|
uint64_t ws1;
|
|
uint64_t ws2;
|
|
double exp_res_1;
|
|
double exp_res_2;
|
|
};
|
|
|
|
TEST(MSA_ffint_u) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseMsa2RF_U_F tc_s[] = {
|
|
{0, 345, 234, 1000, 0.f, 345.f, 234.f, 1000.f}};
|
|
|
|
struct TestCaseMsa2RF_U_D tc_d[] = {{0, 345, 0., 345.},
|
|
{234, 1000, 234., 1000.}};
|
|
|
|
for (size_t i = 0; i < sizeof(tc_s) / sizeof(TestCaseMsa2RF_U_F); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_s[i]),
|
|
[](MacroAssembler& assm) { __ ffint_u_w(w2, w0); });
|
|
}
|
|
for (size_t i = 0; i < sizeof(tc_d) / sizeof(TestCaseMsa2RF_U_D); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]),
|
|
[](MacroAssembler& assm) { __ ffint_u_d(w2, w0); });
|
|
}
|
|
}
|
|
|
|
struct TestCaseMsa2RF_I_F {
|
|
int32_t ws1;
|
|
int32_t ws2;
|
|
int32_t ws3;
|
|
int32_t ws4;
|
|
float exp_res_1;
|
|
float exp_res_2;
|
|
float exp_res_3;
|
|
float exp_res_4;
|
|
};
|
|
|
|
struct TestCaseMsa2RF_I_D {
|
|
int64_t ws1;
|
|
int64_t ws2;
|
|
double exp_res_1;
|
|
double exp_res_2;
|
|
};
|
|
|
|
TEST(MSA_ffint_s) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseMsa2RF_I_F tc_s[] = {
|
|
{0, 345, -234, 1000, 0.f, 345.f, -234.f, 1000.f}};
|
|
|
|
struct TestCaseMsa2RF_I_D tc_d[] = {{0, 345, 0., 345.},
|
|
{-234, 1000, -234., 1000.}};
|
|
|
|
for (size_t i = 0; i < sizeof(tc_s) / sizeof(TestCaseMsa2RF_I_F); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_s[i]),
|
|
[](MacroAssembler& assm) { __ ffint_s_w(w2, w0); });
|
|
}
|
|
for (size_t i = 0; i < sizeof(tc_d) / sizeof(TestCaseMsa2RF_I_D); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]),
|
|
[](MacroAssembler& assm) { __ ffint_s_d(w2, w0); });
|
|
}
|
|
}
|
|
|
|
struct TestCaseMsa2RF_U16_F {
|
|
uint16_t ws1;
|
|
uint16_t ws2;
|
|
uint16_t ws3;
|
|
uint16_t ws4;
|
|
uint16_t ws5;
|
|
uint16_t ws6;
|
|
uint16_t ws7;
|
|
uint16_t ws8;
|
|
float exp_res_1;
|
|
float exp_res_2;
|
|
float exp_res_3;
|
|
float exp_res_4;
|
|
};
|
|
|
|
struct TestCaseMsa2RF_F_D {
|
|
float ws1;
|
|
float ws2;
|
|
float ws3;
|
|
float ws4;
|
|
double exp_res_1;
|
|
double exp_res_2;
|
|
};
|
|
|
|
TEST(MSA_fexupl) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
const float inf_float = std::numeric_limits<float>::infinity();
|
|
const double inf_double = std::numeric_limits<double>::infinity();
|
|
|
|
struct TestCaseMsa2RF_U16_F tc_s[] = {
|
|
{1, 2, 0x7C00, 0x0C00, 0, 0x7C00, 0xFC00, 0x8000, 0.f, inf_float,
|
|
-inf_float, -0.f},
|
|
{0xFC00, 0xFFFF, 0x00FF, 0x8000, 0x81FE, 0x8000, 0x0345, 0xAAAA,
|
|
-3.0398368835e-5f, -0.f, 4.9889088e-5f, -5.2062988281e-2f},
|
|
{3, 4, 0x5555, 6, 0x2AAA, 0x8700, 0x7777, 0x6A8B, 5.2062988281e-2f,
|
|
-1.06811523458e-4f, 3.0576e4f, 3.35e3f}};
|
|
|
|
struct TestCaseMsa2RF_F_D tc_d[] = {
|
|
{0.f, 123.456f, inf_float, -0.f, inf_double, -0.},
|
|
{-inf_float, -3.f, 0.f, -inf_float, 0., -inf_double},
|
|
{2.3f, 3., 1.37747639043129518071e-41f, -3.22084585277826e35f,
|
|
1.37747639043129518071e-41, -3.22084585277826e35}};
|
|
|
|
for (size_t i = 0; i < sizeof(tc_s) / sizeof(TestCaseMsa2RF_U16_F); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_s[i]),
|
|
[](MacroAssembler& assm) { __ fexupl_w(w2, w0); });
|
|
}
|
|
for (size_t i = 0; i < sizeof(tc_d) / sizeof(TestCaseMsa2RF_F_D); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]),
|
|
[](MacroAssembler& assm) { __ fexupl_d(w2, w0); });
|
|
}
|
|
}
|
|
|
|
TEST(MSA_fexupr) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
const float inf_float = std::numeric_limits<float>::infinity();
|
|
const double inf_double = std::numeric_limits<double>::infinity();
|
|
|
|
struct TestCaseMsa2RF_U16_F tc_s[] = {
|
|
{0, 0x7C00, 0xFC00, 0x8000, 1, 2, 0x7C00, 0x0C00, 0.f, inf_float,
|
|
-inf_float, -0.f},
|
|
{0x81FE, 0x8000, 0x0345, 0xAAAA, 0xFC00, 0xFFFF, 0x00FF, 0x8000,
|
|
-3.0398368835e-5f, -0.f, 4.9889088e-5f, -5.2062988281e-2f},
|
|
{0x2AAA, 0x8700, 0x7777, 0x6A8B, 3, 4, 0x5555, 6, 5.2062988281e-2f,
|
|
-1.06811523458e-4f, 3.0576e4f, 3.35e3f}};
|
|
|
|
struct TestCaseMsa2RF_F_D tc_d[] = {
|
|
{inf_float, -0.f, 0.f, 123.456f, inf_double, -0.},
|
|
{0.f, -inf_float, -inf_float, -3.f, 0., -inf_double},
|
|
{1.37747639043129518071e-41f, -3.22084585277826e35f, 2.3f, 3.,
|
|
1.37747639043129518071e-41, -3.22084585277826e35}};
|
|
|
|
for (size_t i = 0; i < sizeof(tc_s) / sizeof(TestCaseMsa2RF_U16_F); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_s[i]),
|
|
[](MacroAssembler& assm) { __ fexupr_w(w2, w0); });
|
|
}
|
|
for (size_t i = 0; i < sizeof(tc_d) / sizeof(TestCaseMsa2RF_F_D); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]),
|
|
[](MacroAssembler& assm) { __ fexupr_d(w2, w0); });
|
|
}
|
|
}
|
|
|
|
struct TestCaseMsa2RF_U32_D {
|
|
uint32_t ws1;
|
|
uint32_t ws2;
|
|
uint32_t ws3;
|
|
uint32_t ws4;
|
|
double exp_res_1;
|
|
double exp_res_2;
|
|
};
|
|
|
|
TEST(MSA_ffql) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseMsa2RF_U16_F tc_s[] = {{0, 3, 0xFFFF, 0x8000, 0x8000, 0xE000,
|
|
0x0FF0, 0, -1.f, -0.25f,
|
|
0.12451171875f, 0.f}};
|
|
|
|
struct TestCaseMsa2RF_U32_D tc_d[] = {
|
|
{0, 45, 0x80000000, 0xE0000000, -1., -0.25},
|
|
{0x28379, 0xAAAA5555, 0x024903D3, 0, 17.853239085525274277e-3, 0.}};
|
|
|
|
for (size_t i = 0; i < sizeof(tc_s) / sizeof(TestCaseMsa2RF_U16_F); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_s[i]),
|
|
[](MacroAssembler& assm) { __ ffql_w(w2, w0); });
|
|
}
|
|
for (size_t i = 0; i < sizeof(tc_d) / sizeof(TestCaseMsa2RF_U32_D); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]),
|
|
[](MacroAssembler& assm) { __ ffql_d(w2, w0); });
|
|
}
|
|
}
|
|
|
|
TEST(MSA_ffqr) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseMsa2RF_U16_F tc_s[] = {{0x8000, 0xE000, 0x0FF0, 0, 0, 3,
|
|
0xFFFF, 0x8000, -1.f, -0.25f,
|
|
0.12451171875f, 0.f}};
|
|
|
|
struct TestCaseMsa2RF_U32_D tc_d[] = {
|
|
{0x80000000, 0xE0000000, 0, 45, -1., -0.25},
|
|
{0x024903D3, 0, 0x28379, 0xAAAA5555, 17.853239085525274277e-3, 0.}};
|
|
|
|
for (size_t i = 0; i < sizeof(tc_s) / sizeof(TestCaseMsa2RF_U16_F); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_s[i]),
|
|
[](MacroAssembler& assm) { __ ffqr_w(w2, w0); });
|
|
}
|
|
for (size_t i = 0; i < sizeof(tc_d) / sizeof(TestCaseMsa2RF_U32_D); ++i) {
|
|
run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]),
|
|
[](MacroAssembler& assm) { __ ffqr_d(w2, w0); });
|
|
}
|
|
}
|
|
|
|
struct TestCaseMsaVector {
|
|
uint64_t wd_lo;
|
|
uint64_t wd_hi;
|
|
uint64_t ws_lo;
|
|
uint64_t ws_hi;
|
|
uint64_t wt_lo;
|
|
uint64_t wt_hi;
|
|
};
|
|
|
|
template <typename InstFunc, typename OperFunc>
|
|
void run_msa_vector(struct TestCaseMsaVector* input,
|
|
InstFunc GenerateVectorInstructionFunc,
|
|
OperFunc GenerateOperationFunc) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
CpuFeatureScope fscope(&assm, MIPS_SIMD);
|
|
msa_reg_t res;
|
|
|
|
load_elements_of_vector(&assm, &(input->ws_lo), w0, t0, t1);
|
|
load_elements_of_vector(&assm, &(input->wt_lo), w2, t0, t1);
|
|
load_elements_of_vector(&assm, &(input->wd_lo), w4, t0, t1);
|
|
|
|
GenerateVectorInstructionFunc(assm);
|
|
|
|
store_elements_of_vector(&assm, w4, a0);
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
#ifdef OBJECT_PRINT
|
|
code->Print(std::cout);
|
|
#endif
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
|
|
(f.Call(&res, 0, 0, 0, 0));
|
|
|
|
CHECK_EQ(GenerateOperationFunc(input->wd_lo, input->ws_lo, input->wt_lo),
|
|
res.d[0]);
|
|
CHECK_EQ(GenerateOperationFunc(input->wd_hi, input->ws_hi, input->wt_hi),
|
|
res.d[1]);
|
|
}
|
|
|
|
TEST(MSA_vector) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseMsaVector tc[] = {
|
|
// wd_lo, wd_hi, ws_lo, ws_hi, wt_lo, wt_hi
|
|
{0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636, 0xDCD39D91F9057627,
|
|
0x64BE4F6DBE9CAA51, 0x6B23DE1A687D9CB9, 0x49547AAD691DA4CA},
|
|
{0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636, 0x401614523D830549,
|
|
0xD7C46D613F50EDDD, 0x52284CBC60A1562B, 0x1756ED510D8849CD},
|
|
{0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636, 0xD6E2D2EBCB40D72F,
|
|
0x13A619AFCE67B079, 0x36CCE284343E40F9, 0xB4E8F44FD148BF7F}};
|
|
|
|
for (size_t i = 0; i < sizeof(tc) / sizeof(TestCaseMsaVector); ++i) {
|
|
run_msa_vector(
|
|
&tc[i], [](MacroAssembler& assm) { __ and_v(w4, w0, w2); },
|
|
[](uint64_t wd, uint64_t ws, uint64_t wt) { return ws & wt; });
|
|
run_msa_vector(
|
|
&tc[i], [](MacroAssembler& assm) { __ or_v(w4, w0, w2); },
|
|
[](uint64_t wd, uint64_t ws, uint64_t wt) { return ws | wt; });
|
|
run_msa_vector(
|
|
&tc[i], [](MacroAssembler& assm) { __ nor_v(w4, w0, w2); },
|
|
[](uint64_t wd, uint64_t ws, uint64_t wt) { return ~(ws | wt); });
|
|
run_msa_vector(
|
|
&tc[i], [](MacroAssembler& assm) { __ xor_v(w4, w0, w2); },
|
|
[](uint64_t wd, uint64_t ws, uint64_t wt) { return ws ^ wt; });
|
|
run_msa_vector(&tc[i], [](MacroAssembler& assm) { __ bmnz_v(w4, w0, w2); },
|
|
[](uint64_t wd, uint64_t ws, uint64_t wt) {
|
|
return (ws & wt) | (wd & ~wt);
|
|
});
|
|
run_msa_vector(&tc[i], [](MacroAssembler& assm) { __ bmz_v(w4, w0, w2); },
|
|
[](uint64_t wd, uint64_t ws, uint64_t wt) {
|
|
return (ws & ~wt) | (wd & wt);
|
|
});
|
|
run_msa_vector(&tc[i], [](MacroAssembler& assm) { __ bsel_v(w4, w0, w2); },
|
|
[](uint64_t wd, uint64_t ws, uint64_t wt) {
|
|
return (ws & ~wd) | (wt & wd);
|
|
});
|
|
}
|
|
}
|
|
|
|
struct TestCaseMsaBit {
|
|
uint64_t wd_lo;
|
|
uint64_t wd_hi;
|
|
uint64_t ws_lo;
|
|
uint64_t ws_hi;
|
|
uint32_t m;
|
|
};
|
|
|
|
template <typename InstFunc, typename OperFunc>
|
|
void run_msa_bit(struct TestCaseMsaBit* input, InstFunc GenerateInstructionFunc,
|
|
OperFunc GenerateOperationFunc) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
CpuFeatureScope fscope(&assm, MIPS_SIMD);
|
|
msa_reg_t res;
|
|
|
|
load_elements_of_vector(&assm, &(input->ws_lo), w0, t0, t1);
|
|
load_elements_of_vector(&assm, &(input->wd_lo), w2, t0, t1);
|
|
|
|
GenerateInstructionFunc(assm, input->m);
|
|
|
|
store_elements_of_vector(&assm, w2, a0);
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
#ifdef OBJECT_PRINT
|
|
code->Print(std::cout);
|
|
#endif
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
|
|
(f.Call(&res, 0, 0, 0, 0));
|
|
|
|
CHECK_EQ(GenerateOperationFunc(input->wd_lo, input->ws_lo, input->m),
|
|
res.d[0]);
|
|
CHECK_EQ(GenerateOperationFunc(input->wd_hi, input->ws_hi, input->m),
|
|
res.d[1]);
|
|
}
|
|
|
|
TEST(MSA_slli_srai_srli) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseMsaBit tc[] = {
|
|
// wd_lo, wd_hi ws_lo, ws_hi, m
|
|
{0, 0, 0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636, 3},
|
|
{0, 0, 0x64BE4F6DBE9CAA51, 0x6B23DE1A687D9CB9, 5},
|
|
{0, 0, 0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 9},
|
|
{0, 0, 0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 13},
|
|
{0, 0, 0x566BE7BA4365B70A, 0x01EBBC1937D76CB4, 21},
|
|
{0, 0, 0x380E2DEB9D3F8AAE, 0x017E0DE0BCC6CA42, 30},
|
|
{0, 0, 0xA46A3A9BCB43F4E5, 0x1C62C8473BDFCFFB, 45},
|
|
{0, 0, 0xF6759D85F23B5A2B, 0x5C042AE42C6D12C1, 61}};
|
|
|
|
#define SLLI_SRLI_DF(lanes, mask, func) \
|
|
[](uint64_t wd, uint64_t ws, uint32_t m) { \
|
|
uint64_t res = 0; \
|
|
int elem_size = kMSARegSize / lanes; \
|
|
for (int i = 0; i < lanes / 2; ++i) { \
|
|
int shift = elem_size * i; \
|
|
uint64_t elem = (ws >> shift) & mask; \
|
|
res |= ((func)&mask) << shift; \
|
|
} \
|
|
return res; \
|
|
}
|
|
|
|
#define SRAI_DF(lanes, mask, func) \
|
|
[](uint64_t wd, uint64_t ws, uint32_t m) { \
|
|
uint64_t res = 0; \
|
|
int elem_size = kMSARegSize / lanes; \
|
|
for (int i = 0; i < lanes / 2; ++i) { \
|
|
int shift = elem_size * i; \
|
|
int64_t elem = \
|
|
static_cast<int64_t>(((ws >> shift) & mask) << (64 - elem_size)) >> \
|
|
(64 - elem_size); \
|
|
res |= static_cast<uint64_t>((func)&mask) << shift; \
|
|
} \
|
|
return res; \
|
|
}
|
|
|
|
for (size_t i = 0; i < sizeof(tc) / sizeof(TestCaseMsaBit); ++i) {
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ slli_b(w2, w0, m % 8); },
|
|
SLLI_SRLI_DF(kMSALanesByte, UINT8_MAX, (elem << (m % elem_size))));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ slli_h(w2, w0, m % 16); },
|
|
SLLI_SRLI_DF(kMSALanesHalf, UINT16_MAX, (elem << (m % elem_size))));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ slli_w(w2, w0, m % 32); },
|
|
SLLI_SRLI_DF(kMSALanesWord, UINT32_MAX, (elem << (m % elem_size))));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ slli_d(w2, w0, m % 64); },
|
|
SLLI_SRLI_DF(kMSALanesDword, UINT64_MAX, (elem << (m % elem_size))));
|
|
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ srli_b(w2, w0, m % 8); },
|
|
SLLI_SRLI_DF(kMSALanesByte, UINT8_MAX, (elem >> (m % elem_size))));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ srli_h(w2, w0, m % 16); },
|
|
SLLI_SRLI_DF(kMSALanesHalf, UINT16_MAX, (elem >> (m % elem_size))));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ srli_w(w2, w0, m % 32); },
|
|
SLLI_SRLI_DF(kMSALanesWord, UINT32_MAX, (elem >> (m % elem_size))));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ srli_d(w2, w0, m % 64); },
|
|
SLLI_SRLI_DF(kMSALanesDword, UINT64_MAX, (elem >> (m % elem_size))));
|
|
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ srlri_b(w2, w0, m % 8); },
|
|
SLLI_SRLI_DF(
|
|
kMSALanesByte, UINT8_MAX,
|
|
(elem >> (m % elem_size)) + ((elem >> (m % elem_size - 1)) & 0x1)));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ srlri_h(w2, w0, m % 16); },
|
|
SLLI_SRLI_DF(
|
|
kMSALanesHalf, UINT16_MAX,
|
|
(elem >> (m % elem_size)) + ((elem >> (m % elem_size - 1)) & 0x1)));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ srlri_w(w2, w0, m % 32); },
|
|
SLLI_SRLI_DF(
|
|
kMSALanesWord, UINT32_MAX,
|
|
(elem >> (m % elem_size)) + ((elem >> (m % elem_size - 1)) & 0x1)));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ srlri_d(w2, w0, m % 64); },
|
|
SLLI_SRLI_DF(
|
|
kMSALanesDword, UINT64_MAX,
|
|
(elem >> (m % elem_size)) + ((elem >> (m % elem_size - 1)) & 0x1)));
|
|
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ srai_b(w2, w0, m % 8); },
|
|
SRAI_DF(kMSALanesByte, UINT8_MAX,
|
|
ArithmeticShiftRight(elem, m % elem_size)));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ srai_h(w2, w0, m % 16); },
|
|
SRAI_DF(kMSALanesHalf, UINT16_MAX,
|
|
ArithmeticShiftRight(elem, m % elem_size)));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ srai_w(w2, w0, m % 32); },
|
|
SRAI_DF(kMSALanesWord, UINT32_MAX,
|
|
ArithmeticShiftRight(elem, m % elem_size)));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ srai_d(w2, w0, m % 64); },
|
|
SRAI_DF(kMSALanesDword, UINT64_MAX,
|
|
ArithmeticShiftRight(elem, m % elem_size)));
|
|
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ srari_b(w2, w0, m % 8); },
|
|
SRAI_DF(kMSALanesByte, UINT8_MAX,
|
|
ArithmeticShiftRight(elem, m % elem_size) +
|
|
((elem >> (m % elem_size - 1)) & 0x1)));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ srari_h(w2, w0, m % 16); },
|
|
SRAI_DF(kMSALanesHalf, UINT16_MAX,
|
|
ArithmeticShiftRight(elem, m % elem_size) +
|
|
((elem >> (m % elem_size - 1)) & 0x1)));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ srari_w(w2, w0, m % 32); },
|
|
SRAI_DF(kMSALanesWord, UINT32_MAX,
|
|
ArithmeticShiftRight(elem, m % elem_size) +
|
|
((elem >> (m % elem_size - 1)) & 0x1)));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ srari_d(w2, w0, m % 64); },
|
|
SRAI_DF(kMSALanesDword, UINT64_MAX,
|
|
ArithmeticShiftRight(elem, m % elem_size) +
|
|
((elem >> (m % elem_size - 1)) & 0x1)));
|
|
}
|
|
#undef SLLI_SRLI_DF
|
|
#undef SRAI_DF
|
|
}
|
|
|
|
TEST(MSA_bclri_bseti_bnegi) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseMsaBit tc[] = {
|
|
// wd_lo, wd_hi, ws_lo, ws_hi, m
|
|
{0, 0, 0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636, 3},
|
|
{0, 0, 0x64BE4F6DBE9CAA51, 0x6B23DE1A687D9CB9, 5},
|
|
{0, 0, 0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 9},
|
|
{0, 0, 0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 13},
|
|
{0, 0, 0x566BE7BA4365B70A, 0x01EBBC1937D76CB4, 21},
|
|
{0, 0, 0x380E2DEB9D3F8AAE, 0x017E0DE0BCC6CA42, 30},
|
|
{0, 0, 0xA46A3A9BCB43F4E5, 0x1C62C8473BDFCFFB, 45},
|
|
{0, 0, 0xF6759D85F23B5A2B, 0x5C042AE42C6D12C1, 61}};
|
|
|
|
#define BCLRI_BSETI_BNEGI_DF(lanes, mask, func) \
|
|
[](uint64_t wd, uint64_t ws, uint32_t m) { \
|
|
uint64_t res = 0; \
|
|
int elem_size = kMSARegSize / lanes; \
|
|
for (int i = 0; i < lanes / 2; ++i) { \
|
|
int shift = elem_size * i; \
|
|
uint64_t elem = (ws >> shift) & mask; \
|
|
res |= ((func)&mask) << shift; \
|
|
} \
|
|
return res; \
|
|
}
|
|
|
|
for (size_t i = 0; i < sizeof(tc) / sizeof(TestCaseMsaBit); ++i) {
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ bclri_b(w2, w0, m % 8); },
|
|
BCLRI_BSETI_BNEGI_DF(kMSALanesByte, UINT8_MAX,
|
|
(~(1ull << (m % elem_size)) & elem)));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ bclri_h(w2, w0, m % 16); },
|
|
BCLRI_BSETI_BNEGI_DF(kMSALanesHalf, UINT16_MAX,
|
|
(~(1ull << (m % elem_size)) & elem)));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ bclri_w(w2, w0, m % 32); },
|
|
BCLRI_BSETI_BNEGI_DF(kMSALanesWord, UINT32_MAX,
|
|
(~(1ull << (m % elem_size)) & elem)));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ bclri_d(w2, w0, m % 64); },
|
|
BCLRI_BSETI_BNEGI_DF(kMSALanesDword, UINT64_MAX,
|
|
(~(1ull << (m % elem_size)) & elem)));
|
|
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ bseti_b(w2, w0, m % 8); },
|
|
BCLRI_BSETI_BNEGI_DF(kMSALanesByte, UINT8_MAX,
|
|
((1ull << (m % elem_size)) | elem)));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ bseti_h(w2, w0, m % 16); },
|
|
BCLRI_BSETI_BNEGI_DF(kMSALanesHalf, UINT16_MAX,
|
|
((1ull << (m % elem_size)) | elem)));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ bseti_w(w2, w0, m % 32); },
|
|
BCLRI_BSETI_BNEGI_DF(kMSALanesWord, UINT32_MAX,
|
|
((1ull << (m % elem_size)) | elem)));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ bseti_d(w2, w0, m % 64); },
|
|
BCLRI_BSETI_BNEGI_DF(kMSALanesDword, UINT64_MAX,
|
|
((1ull << (m % elem_size)) | elem)));
|
|
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ bnegi_b(w2, w0, m % 8); },
|
|
BCLRI_BSETI_BNEGI_DF(kMSALanesByte, UINT8_MAX,
|
|
((1ull << (m % elem_size)) ^ elem)));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ bnegi_h(w2, w0, m % 16); },
|
|
BCLRI_BSETI_BNEGI_DF(kMSALanesHalf, UINT16_MAX,
|
|
((1ull << (m % elem_size)) ^ elem)));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ bnegi_w(w2, w0, m % 32); },
|
|
BCLRI_BSETI_BNEGI_DF(kMSALanesWord, UINT32_MAX,
|
|
((1ull << (m % elem_size)) ^ elem)));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ bnegi_d(w2, w0, m % 64); },
|
|
BCLRI_BSETI_BNEGI_DF(kMSALanesDword, UINT64_MAX,
|
|
((1ull << (m % elem_size)) ^ elem)));
|
|
}
|
|
#undef BCLRI_BSETI_BNEGI_DF
|
|
}
|
|
|
|
TEST(MSA_binsli_binsri) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseMsaBit tc[] = {// wd_lo, wd_hi, ws_lo, ws_hi, m
|
|
{0x53F4457553BBD5B4, 0x5FB8250EACC296B2,
|
|
0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636, 3},
|
|
{0xF61BFDB0F312E6FC, 0xC9437568DD1EA925,
|
|
0x64BE4F6DBE9CAA51, 0x6B23DE1A687D9CB9, 5},
|
|
{0x53F4457553BBD5B4, 0x5FB8250EACC296B2,
|
|
0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 9},
|
|
{0xF61BFDB0F312E6FC, 0xC9437568DD1EA925,
|
|
0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 13},
|
|
{0x53F4457553BBD5B4, 0x5FB8250EACC296B2,
|
|
0x566BE7BA4365B70A, 0x01EBBC1937D76CB4, 21},
|
|
{0xF61BFDB0F312E6FC, 0xC9437568DD1EA925,
|
|
0x380E2DEB9D3F8AAE, 0x017E0DE0BCC6CA42, 30},
|
|
{0x53F4457553BBD5B4, 0x5FB8250EACC296B2,
|
|
0xA46A3A9BCB43F4E5, 0x1C62C8473BDFCFFB, 45},
|
|
{0xF61BFDB0F312E6FC, 0xC9437568DD1EA925,
|
|
0xF6759D85F23B5A2B, 0x5C042AE42C6D12C1, 61}};
|
|
|
|
#define BINSLI_BINSRI_DF(lanes, mask, func) \
|
|
[](uint64_t wd, uint64_t ws, uint32_t m) { \
|
|
uint64_t res = 0; \
|
|
int elem_size = kMSARegSize / lanes; \
|
|
int bits = m % elem_size + 1; \
|
|
for (int i = 0; i < lanes / 2; ++i) { \
|
|
int shift = elem_size * i; \
|
|
uint64_t ws_elem = (ws >> shift) & mask; \
|
|
if (bits == elem_size) { \
|
|
res |= (ws_elem & mask) << shift; \
|
|
} else { \
|
|
uint64_t r_mask = (1ull << bits) - 1; \
|
|
uint64_t l_mask = r_mask << (elem_size - bits); \
|
|
USE(l_mask); \
|
|
uint64_t wd_elem = (wd >> shift) & mask; \
|
|
res |= ((func)&mask) << shift; \
|
|
} \
|
|
} \
|
|
return res; \
|
|
}
|
|
|
|
for (size_t i = 0; i < sizeof(tc) / sizeof(TestCaseMsaBit); ++i) {
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ binsli_b(w2, w0, m % 8); },
|
|
BINSLI_BINSRI_DF(kMSALanesByte, UINT8_MAX,
|
|
((ws_elem & l_mask) | (wd_elem & ~l_mask))));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ binsli_h(w2, w0, m % 16); },
|
|
BINSLI_BINSRI_DF(kMSALanesHalf, UINT16_MAX,
|
|
((ws_elem & l_mask) | (wd_elem & ~l_mask))));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ binsli_w(w2, w0, m % 32); },
|
|
BINSLI_BINSRI_DF(kMSALanesWord, UINT32_MAX,
|
|
((ws_elem & l_mask) | (wd_elem & ~l_mask))));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ binsli_d(w2, w0, m % 64); },
|
|
BINSLI_BINSRI_DF(kMSALanesDword, UINT64_MAX,
|
|
((ws_elem & l_mask) | (wd_elem & ~l_mask))));
|
|
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ binsri_b(w2, w0, m % 8); },
|
|
BINSLI_BINSRI_DF(kMSALanesByte, UINT8_MAX,
|
|
((ws_elem & r_mask) | (wd_elem & ~r_mask))));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ binsri_h(w2, w0, m % 16); },
|
|
BINSLI_BINSRI_DF(kMSALanesHalf, UINT16_MAX,
|
|
((ws_elem & r_mask) | (wd_elem & ~r_mask))));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ binsri_w(w2, w0, m % 32); },
|
|
BINSLI_BINSRI_DF(kMSALanesWord, UINT32_MAX,
|
|
((ws_elem & r_mask) | (wd_elem & ~r_mask))));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ binsri_d(w2, w0, m % 64); },
|
|
BINSLI_BINSRI_DF(kMSALanesDword, UINT64_MAX,
|
|
((ws_elem & r_mask) | (wd_elem & ~r_mask))));
|
|
}
|
|
#undef BINSLI_BINSRI_DF
|
|
}
|
|
|
|
TEST(MSA_sat_s_sat_u) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseMsaBit tc[] = {
|
|
// wd_lo, wd_hi, ws_lo, ws_hi, m
|
|
{0, 0, 0xF35862E13E3808B0, 0x4F41FFDEF2BFE636, 3},
|
|
{0, 0, 0x64BE4F6DBE9CAA51, 0x6B23DE1A687D9CB9, 5},
|
|
{0, 0, 0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 9},
|
|
{0, 0, 0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 13},
|
|
{0, 0, 0x566BE7BA4365B70A, 0x01EBBC1937D76CB4, 21},
|
|
{0, 0, 0x380E2DEB9D3F8AAE, 0x017E0DE0BCC6CA42, 30},
|
|
{0, 0, 0xA46A3A9BCB43F4E5, 0x1C62C8473BDFCFFB, 45},
|
|
{0, 0, 0xF6759D85F23B5A2B, 0x5C042AE42C6D12C1, 61}};
|
|
|
|
#define SAT_DF(lanes, mask, func) \
|
|
[](uint64_t wd, uint64_t ws, uint32_t m) { \
|
|
uint64_t res = 0; \
|
|
int elem_size = kMSARegSize / lanes; \
|
|
m %= elem_size; \
|
|
for (int i = 0; i < lanes / 2; ++i) { \
|
|
int shift = elem_size * i; \
|
|
uint64_t elem_u64 = (ws >> shift) & mask; \
|
|
int64_t elem_i64 = static_cast<int64_t>(elem_u64 << (64 - elem_size)) >> \
|
|
(64 - elem_size); \
|
|
USE(elem_i64); \
|
|
res |= ((func)&mask) << shift; \
|
|
} \
|
|
return res; \
|
|
}
|
|
|
|
#define M_MAX_INT(x) static_cast<int64_t>((1LL << ((x)-1)) - 1)
|
|
#define M_MIN_INT(x) static_cast<int64_t>(-(1LL << ((x)-1)))
|
|
#define M_MAX_UINT(x) static_cast<uint64_t>(-1ULL >> (64 - (x)))
|
|
|
|
for (size_t i = 0; i < sizeof(tc) / sizeof(TestCaseMsaBit); ++i) {
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ sat_u_b(w2, w0, m % 8); },
|
|
SAT_DF(kMSALanesByte, UINT8_MAX,
|
|
(elem_u64 < M_MAX_UINT(m + 1) ? elem_u64 : M_MAX_UINT(m + 1))));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ sat_u_h(w2, w0, m % 16); },
|
|
SAT_DF(kMSALanesHalf, UINT16_MAX,
|
|
(elem_u64 < M_MAX_UINT(m + 1) ? elem_u64 : M_MAX_UINT(m + 1))));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ sat_u_w(w2, w0, m % 32); },
|
|
SAT_DF(kMSALanesWord, UINT32_MAX,
|
|
(elem_u64 < M_MAX_UINT(m + 1) ? elem_u64 : M_MAX_UINT(m + 1))));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ sat_u_d(w2, w0, m % 64); },
|
|
SAT_DF(kMSALanesDword, UINT64_MAX,
|
|
(elem_u64 < M_MAX_UINT(m + 1) ? elem_u64 : M_MAX_UINT(m + 1))));
|
|
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ sat_s_b(w2, w0, m % 8); },
|
|
SAT_DF(
|
|
kMSALanesByte, UINT8_MAX,
|
|
(elem_i64 < M_MIN_INT(m + 1)
|
|
? M_MIN_INT(m + 1)
|
|
: elem_i64 > M_MAX_INT(m + 1) ? M_MAX_INT(m + 1) : elem_i64)));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ sat_s_h(w2, w0, m % 16); },
|
|
SAT_DF(
|
|
kMSALanesHalf, UINT16_MAX,
|
|
(elem_i64 < M_MIN_INT(m + 1)
|
|
? M_MIN_INT(m + 1)
|
|
: elem_i64 > M_MAX_INT(m + 1) ? M_MAX_INT(m + 1) : elem_i64)));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ sat_s_w(w2, w0, m % 32); },
|
|
SAT_DF(
|
|
kMSALanesWord, UINT32_MAX,
|
|
(elem_i64 < M_MIN_INT(m + 1)
|
|
? M_MIN_INT(m + 1)
|
|
: elem_i64 > M_MAX_INT(m + 1) ? M_MAX_INT(m + 1) : elem_i64)));
|
|
run_msa_bit(
|
|
&tc[i],
|
|
[](MacroAssembler& assm, uint32_t m) { __ sat_s_d(w2, w0, m % 64); },
|
|
SAT_DF(
|
|
kMSALanesDword, UINT64_MAX,
|
|
(elem_i64 < M_MIN_INT(m + 1)
|
|
? M_MIN_INT(m + 1)
|
|
: elem_i64 > M_MAX_INT(m + 1) ? M_MAX_INT(m + 1) : elem_i64)));
|
|
}
|
|
|
|
#undef SAT_DF
|
|
#undef M_MAX_INT
|
|
#undef M_MIN_INT
|
|
#undef M_MAX_UINT
|
|
}
|
|
|
|
template <typename InstFunc, typename OperFunc>
|
|
void run_msa_i10(int32_t input, InstFunc GenerateVectorInstructionFunc,
|
|
OperFunc GenerateOperationFunc) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
CpuFeatureScope fscope(&assm, MIPS_SIMD);
|
|
msa_reg_t res;
|
|
|
|
GenerateVectorInstructionFunc(assm, input);
|
|
|
|
store_elements_of_vector(&assm, w0, a0);
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
#ifdef OBJECT_PRINT
|
|
code->Print(std::cout);
|
|
#endif
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
|
|
(f.Call(&res, 0, 0, 0, 0));
|
|
|
|
CHECK_EQ(GenerateOperationFunc(input), res.d[0]);
|
|
CHECK_EQ(GenerateOperationFunc(input), res.d[1]);
|
|
}
|
|
|
|
TEST(MSA_ldi) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
// signed 10bit integers: -512 .. 511
|
|
int32_t tc[] = {0, -1, 1, 256, -256, -178, 352, -512, 511};
|
|
|
|
#define LDI_DF(lanes, mask) \
|
|
[](int32_t s10) { \
|
|
uint64_t res = 0; \
|
|
int elem_size = kMSARegSize / lanes; \
|
|
int64_t s10_64 = \
|
|
ArithmeticShiftRight(static_cast<int64_t>(s10) << 54, 54); \
|
|
for (int i = 0; i < lanes / 2; ++i) { \
|
|
int shift = elem_size * i; \
|
|
res |= static_cast<uint64_t>(s10_64 & mask) << shift; \
|
|
} \
|
|
return res; \
|
|
}
|
|
|
|
for (size_t i = 0; i < sizeof(tc) / sizeof(int32_t); ++i) {
|
|
run_msa_i10(tc[i],
|
|
[](MacroAssembler& assm, int32_t s10) { __ ldi_b(w0, s10); },
|
|
LDI_DF(kMSALanesByte, UINT8_MAX));
|
|
run_msa_i10(tc[i],
|
|
[](MacroAssembler& assm, int32_t s10) { __ ldi_h(w0, s10); },
|
|
LDI_DF(kMSALanesHalf, UINT16_MAX));
|
|
run_msa_i10(tc[i],
|
|
[](MacroAssembler& assm, int32_t s10) { __ ldi_w(w0, s10); },
|
|
LDI_DF(kMSALanesWord, UINT32_MAX));
|
|
run_msa_i10(tc[i],
|
|
[](MacroAssembler& assm, int32_t s10) { __ ldi_d(w0, s10); },
|
|
LDI_DF(kMSALanesDword, UINT64_MAX));
|
|
}
|
|
#undef LDI_DF
|
|
}
|
|
|
|
template <typename T, typename InstFunc>
|
|
void run_msa_mi10(InstFunc GenerateVectorInstructionFunc) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
CpuFeatureScope fscope(&assm, MIPS_SIMD);
|
|
T in_test_vector[1024];
|
|
T out_test_vector[1024];
|
|
|
|
T* in_array_middle = in_test_vector + arraysize(in_test_vector) / 2;
|
|
T* out_array_middle = out_test_vector + arraysize(out_test_vector) / 2;
|
|
|
|
v8::base::RandomNumberGenerator rand_gen(FLAG_random_seed);
|
|
for (unsigned int i = 0; i < arraysize(in_test_vector); i++) {
|
|
in_test_vector[i] = static_cast<T>(rand_gen.NextInt());
|
|
out_test_vector[i] = 0;
|
|
}
|
|
|
|
GenerateVectorInstructionFunc(assm);
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
#ifdef OBJECT_PRINT
|
|
code->Print(std::cout);
|
|
#endif
|
|
auto f = GeneratedCode<F4>::FromCode(*code);
|
|
|
|
(f.Call(in_array_middle, out_array_middle, 0, 0, 0));
|
|
|
|
CHECK_EQ(memcmp(in_test_vector, out_test_vector, arraysize(in_test_vector)),
|
|
0);
|
|
}
|
|
|
|
TEST(MSA_load_store_vector) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
run_msa_mi10<uint8_t>([](MacroAssembler& assm) {
|
|
for (int i = -512; i < 512; i += 16) {
|
|
__ ld_b(w0, MemOperand(a0, i));
|
|
__ st_b(w0, MemOperand(a1, i));
|
|
}
|
|
});
|
|
run_msa_mi10<uint16_t>([](MacroAssembler& assm) {
|
|
for (int i = -512; i < 512; i += 8) {
|
|
__ ld_h(w0, MemOperand(a0, i));
|
|
__ st_h(w0, MemOperand(a1, i));
|
|
}
|
|
});
|
|
run_msa_mi10<uint32_t>([](MacroAssembler& assm) {
|
|
for (int i = -512; i < 512; i += 4) {
|
|
__ ld_w(w0, MemOperand(a0, i));
|
|
__ st_w(w0, MemOperand(a1, i));
|
|
}
|
|
});
|
|
run_msa_mi10<uint64_t>([](MacroAssembler& assm) {
|
|
for (int i = -512; i < 512; i += 2) {
|
|
__ ld_d(w0, MemOperand(a0, i));
|
|
__ st_d(w0, MemOperand(a1, i));
|
|
}
|
|
});
|
|
}
|
|
|
|
struct TestCaseMsa3R {
|
|
uint64_t ws_lo;
|
|
uint64_t ws_hi;
|
|
uint64_t wt_lo;
|
|
uint64_t wt_hi;
|
|
uint64_t wd_lo;
|
|
uint64_t wd_hi;
|
|
};
|
|
|
|
static const uint64_t Unpredictable = 0x312014017725ll;
|
|
|
|
template <typename InstFunc, typename OperFunc>
|
|
void run_msa_3r(struct TestCaseMsa3R* input, InstFunc GenerateI5InstructionFunc,
|
|
OperFunc GenerateOperationFunc) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
CpuFeatureScope fscope(&assm, MIPS_SIMD);
|
|
msa_reg_t res;
|
|
|
|
load_elements_of_vector(&assm, &(input->wt_lo), w0, t0, t1);
|
|
load_elements_of_vector(&assm, &(input->ws_lo), w1, t0, t1);
|
|
load_elements_of_vector(&assm, &(input->wd_lo), w2, t0, t1);
|
|
|
|
GenerateI5InstructionFunc(assm);
|
|
|
|
store_elements_of_vector(&assm, w2, a0);
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
#ifdef OBJECT_PRINT
|
|
code->Print(std::cout);
|
|
#endif
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
|
|
(f.Call(&res, 0, 0, 0, 0));
|
|
|
|
GenerateOperationFunc(&input->ws_lo, &input->wt_lo, &input->wd_lo);
|
|
if (input->wd_lo != Unpredictable) {
|
|
CHECK_EQ(input->wd_lo, res.d[0]);
|
|
}
|
|
if (input->wd_hi != Unpredictable) {
|
|
CHECK_EQ(input->wd_hi, res.d[1]);
|
|
}
|
|
}
|
|
|
|
TEST(MSA_3R_instructions) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
struct TestCaseMsa3R tc[] = {
|
|
{0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0x1169751BB9A7D9C3,
|
|
0xF7A594AEC8EF8A9C, 0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C},
|
|
{0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0x2B665362C4E812DF,
|
|
0x3A0D80D68B3F8BC8, 0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8},
|
|
{0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0x1169751BB9A7D9C3,
|
|
0xF7A594AEC8EF8A9C, 0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C},
|
|
{0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0x2B665362C4E812DF,
|
|
0x3A0D80D68B3F8BC8, 0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8},
|
|
{0xFFAB807F807FFFCD, 0x7F23FF80FF567F80, 0xFFAB807F807FFFCD,
|
|
0x7F23FF80FF567F80, 0xFFAB807F807FFFCD, 0x7F23FF80FF567F80},
|
|
{0x80FFEFFF7F12807F, 0x807F80FF7FDEFF78, 0x80FFEFFF7F12807F,
|
|
0x807F80FF7FDEFF78, 0x80FFEFFF7F12807F, 0x807F80FF7FDEFF78},
|
|
{0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF,
|
|
0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF},
|
|
{0x0000000000000000, 0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF,
|
|
0x0000000000000000, 0x0000000000000000, 0xFFFFFFFFFFFFFFFF},
|
|
{0xFFFF0000FFFF0000, 0xFFFF0000FFFF0000, 0xFFFF0000FFFF0000,
|
|
0xFFFF0000FFFF0000, 0xFFFF0000FFFF0000, 0xFFFF0000FFFF0000},
|
|
{0xFF00FF00FF00FF00, 0xFF00FF00FF00FF00, 0xFF00FF00FF00FF00,
|
|
0xFF00FF00FF00FF00, 0xFF00FF00FF00FF00, 0xFF00FF00FF00FF00},
|
|
{0xF0F0F0F0F0F0F0F0, 0xF0F0F0F0F0F0F0F0, 0xF0F0F0F0F0F0F0F0,
|
|
0xF0F0F0F0F0F0F0F0, 0xF0F0F0F0F0F0F0F0, 0xF0F0F0F0F0F0F0F0},
|
|
{0xFF0000FFFF0000FF, 0xFF0000FFFF0000FF, 0xFF0000FFFF0000FF,
|
|
0xFF0000FFFF0000FF, 0xFF0000FFFF0000FF, 0xFF0000FFFF0000FF},
|
|
{0xFFFF00000000FFFF, 0xFFFF00000000FFFF, 0xFFFF00000000FFFF,
|
|
0xFFFF00000000FFFF, 0xFFFF00000000FFFF, 0xFFFF00000000FFFF}};
|
|
|
|
#define SLL_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T src_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T shift_op = static_cast<T>((wt[i] >> shift) & mask) % size_in_bits; \
|
|
res |= (static_cast<uint64_t>(src_op << shift_op) & mask) << shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define SRA_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T src_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T shift_op = ((wt[i] >> shift) & mask) % size_in_bits; \
|
|
res |= (static_cast<uint64_t>(ArithmeticShiftRight(src_op, shift_op) & \
|
|
mask)) \
|
|
<< shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define SRL_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T src_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T shift_op = static_cast<T>(((wt[i] >> shift) & mask) % size_in_bits); \
|
|
res |= (static_cast<uint64_t>(src_op >> shift_op) & mask) << shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define BCRL_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T src_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T shift_op = static_cast<T>(((wt[i] >> shift) & mask) % size_in_bits); \
|
|
T r = (static_cast<T>(~(1ull << shift_op)) & src_op) & mask; \
|
|
res |= static_cast<uint64_t>(r) << shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define BSET_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T src_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T shift_op = static_cast<T>(((wt[i] >> shift) & mask) % size_in_bits); \
|
|
T r = (static_cast<T>(1ull << shift_op) | src_op) & mask; \
|
|
res |= static_cast<uint64_t>(r) << shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define BNEG_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T src_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T shift_op = static_cast<T>(((wt[i] >> shift) & mask) % size_in_bits); \
|
|
T r = (static_cast<T>(1ull << shift_op) ^ src_op) & mask; \
|
|
res |= static_cast<uint64_t>(r) << shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define BINSL_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T ws_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T wd_op = static_cast<T>((wd[i] >> shift) & mask); \
|
|
T shift_op = static_cast<T>(((wt[i] >> shift) & mask) % size_in_bits); \
|
|
int bits = shift_op + 1; \
|
|
T r; \
|
|
if (bits == size_in_bits) { \
|
|
r = static_cast<T>(ws_op); \
|
|
} else { \
|
|
uint64_t mask2 = ((1ull << bits) - 1) << (size_in_bits - bits); \
|
|
r = static_cast<T>((static_cast<T>(mask2) & ws_op) | \
|
|
(static_cast<T>(~mask2) & wd_op)); \
|
|
} \
|
|
res |= static_cast<uint64_t>(r) << shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define BINSR_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T ws_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T wd_op = static_cast<T>((wd[i] >> shift) & mask); \
|
|
T shift_op = static_cast<T>(((wt[i] >> shift) & mask) % size_in_bits); \
|
|
int bits = shift_op + 1; \
|
|
T r; \
|
|
if (bits == size_in_bits) { \
|
|
r = static_cast<T>(ws_op); \
|
|
} else { \
|
|
uint64_t mask2 = (1ull << bits) - 1; \
|
|
r = static_cast<T>((static_cast<T>(mask2) & ws_op) | \
|
|
(static_cast<T>(~mask2) & wd_op)); \
|
|
} \
|
|
res |= static_cast<uint64_t>(r) << shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define ADDV_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T ws_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T wt_op = static_cast<T>((wt[i] >> shift) & mask); \
|
|
res |= (static_cast<uint64_t>(ws_op + wt_op) & mask) << shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define SUBV_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T ws_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T wt_op = static_cast<T>((wt[i] >> shift) & mask); \
|
|
res |= (static_cast<uint64_t>(ws_op - wt_op) & mask) << shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define MAX_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T ws_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T wt_op = static_cast<T>((wt[i] >> shift) & mask); \
|
|
res |= (static_cast<uint64_t>(Max<T>(ws_op, wt_op)) & mask) << shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define MIN_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T ws_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T wt_op = static_cast<T>((wt[i] >> shift) & mask); \
|
|
res |= (static_cast<uint64_t>(Min<T>(ws_op, wt_op)) & mask) << shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define MAXA_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T ws_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T wt_op = static_cast<T>((wt[i] >> shift) & mask); \
|
|
res |= \
|
|
(static_cast<uint64_t>(Nabs(ws_op) < Nabs(wt_op) ? ws_op : wt_op) & \
|
|
mask) \
|
|
<< shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define MINA_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T ws_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T wt_op = static_cast<T>((wt[i] >> shift) & mask); \
|
|
res |= \
|
|
(static_cast<uint64_t>(Nabs(ws_op) > Nabs(wt_op) ? ws_op : wt_op) & \
|
|
mask) \
|
|
<< shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define CEQ_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T ws_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T wt_op = static_cast<T>((wt[i] >> shift) & mask); \
|
|
res |= (static_cast<uint64_t>(!Compare(ws_op, wt_op) ? -1ull : 0ull) & \
|
|
mask) \
|
|
<< shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define CLT_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T ws_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T wt_op = static_cast<T>((wt[i] >> shift) & mask); \
|
|
res |= (static_cast<uint64_t>((Compare(ws_op, wt_op) == -1) ? -1ull \
|
|
: 0ull) & \
|
|
mask) \
|
|
<< shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define CLE_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T ws_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T wt_op = static_cast<T>((wt[i] >> shift) & mask); \
|
|
res |= (static_cast<uint64_t>((Compare(ws_op, wt_op) != 1) ? -1ull \
|
|
: 0ull) & \
|
|
mask) \
|
|
<< shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define ADD_A_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T ws_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T wt_op = static_cast<T>((wt[i] >> shift) & mask); \
|
|
res |= (static_cast<uint64_t>(Abs(ws_op) + Abs(wt_op)) & mask) << shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define ADDS_A_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T ws_op = Nabs(static_cast<T>((ws[i] >> shift) & mask)); \
|
|
T wt_op = Nabs(static_cast<T>((wt[i] >> shift) & mask)); \
|
|
T r; \
|
|
if (ws_op < -std::numeric_limits<T>::max() - wt_op) { \
|
|
r = std::numeric_limits<T>::max(); \
|
|
} else { \
|
|
r = -(ws_op + wt_op); \
|
|
} \
|
|
res |= (static_cast<uint64_t>(r) & mask) << shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define ADDS_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T ws_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T wt_op = static_cast<T>((wt[i] >> shift) & mask); \
|
|
res |= (static_cast<uint64_t>(SaturateAdd(ws_op, wt_op)) & mask) \
|
|
<< shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define AVE_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T ws_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T wt_op = static_cast<T>((wt[i] >> shift) & mask); \
|
|
res |= (static_cast<uint64_t>( \
|
|
((wt_op & ws_op) + ((ws_op ^ wt_op) >> 1)) & mask)) \
|
|
<< shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define AVER_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T ws_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T wt_op = static_cast<T>((wt[i] >> shift) & mask); \
|
|
res |= (static_cast<uint64_t>( \
|
|
((wt_op | ws_op) - ((ws_op ^ wt_op) >> 1)) & mask)) \
|
|
<< shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define SUBS_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T ws_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T wt_op = static_cast<T>((wt[i] >> shift) & mask); \
|
|
res |= (static_cast<uint64_t>(SaturateSub(ws_op, wt_op)) & mask) \
|
|
<< shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define SUBSUS_U_DF(T, lanes, mask) \
|
|
using uT = typename std::make_unsigned<T>::type; \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
uT ws_op = static_cast<uT>((ws[i] >> shift) & mask); \
|
|
T wt_op = static_cast<T>((wt[i] >> shift) & mask); \
|
|
T r; \
|
|
if (wt_op > 0) { \
|
|
uT wtu = static_cast<uT>(wt_op); \
|
|
if (wtu > ws_op) { \
|
|
r = 0; \
|
|
} else { \
|
|
r = static_cast<T>(ws_op - wtu); \
|
|
} \
|
|
} else { \
|
|
if (ws_op > std::numeric_limits<uT>::max() + wt_op) { \
|
|
r = static_cast<T>(std::numeric_limits<uT>::max()); \
|
|
} else { \
|
|
r = static_cast<T>(ws_op - wt_op); \
|
|
} \
|
|
} \
|
|
res |= (static_cast<uint64_t>(r) & mask) << shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define SUBSUU_S_DF(T, lanes, mask) \
|
|
using uT = typename std::make_unsigned<T>::type; \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
uT ws_op = static_cast<uT>((ws[i] >> shift) & mask); \
|
|
uT wt_op = static_cast<uT>((wt[i] >> shift) & mask); \
|
|
uT wdu; \
|
|
T r; \
|
|
if (ws_op > wt_op) { \
|
|
wdu = ws_op - wt_op; \
|
|
if (wdu > std::numeric_limits<T>::max()) { \
|
|
r = std::numeric_limits<T>::max(); \
|
|
} else { \
|
|
r = static_cast<T>(wdu); \
|
|
} \
|
|
} else { \
|
|
wdu = wt_op - ws_op; \
|
|
CHECK(-std::numeric_limits<T>::max() == \
|
|
std::numeric_limits<T>::min() + 1); \
|
|
if (wdu <= std::numeric_limits<T>::max()) { \
|
|
r = -static_cast<T>(wdu); \
|
|
} else { \
|
|
r = std::numeric_limits<T>::min(); \
|
|
} \
|
|
} \
|
|
res |= (static_cast<uint64_t>(r) & mask) << shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define ASUB_S_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T ws_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T wt_op = static_cast<T>((wt[i] >> shift) & mask); \
|
|
res |= (static_cast<uint64_t>(Abs(ws_op - wt_op)) & mask) << shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define ASUB_U_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T ws_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T wt_op = static_cast<T>((wt[i] >> shift) & mask); \
|
|
res |= (static_cast<uint64_t>(ws_op > wt_op ? ws_op - wt_op \
|
|
: wt_op - ws_op) & \
|
|
mask) \
|
|
<< shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define MULV_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T ws_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T wt_op = static_cast<T>((wt[i] >> shift) & mask); \
|
|
res |= (static_cast<uint64_t>(ws_op * wt_op) & mask) << shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define MADDV_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T ws_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T wt_op = static_cast<T>((wt[i] >> shift) & mask); \
|
|
T wd_op = static_cast<T>((wd[i] >> shift) & mask); \
|
|
res |= (static_cast<uint64_t>(wd_op + ws_op * wt_op) & mask) << shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define MSUBV_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T ws_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T wt_op = static_cast<T>((wt[i] >> shift) & mask); \
|
|
T wd_op = static_cast<T>((wd[i] >> shift) & mask); \
|
|
res |= (static_cast<uint64_t>(wd_op - ws_op * wt_op) & mask) << shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define DIV_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T ws_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T wt_op = static_cast<T>((wt[i] >> shift) & mask); \
|
|
if (wt_op == 0) { \
|
|
res = Unpredictable; \
|
|
break; \
|
|
} \
|
|
res |= (static_cast<uint64_t>(ws_op / wt_op) & mask) << shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define MOD_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T ws_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T wt_op = static_cast<T>((wt[i] >> shift) & mask); \
|
|
if (wt_op == 0) { \
|
|
res = Unpredictable; \
|
|
break; \
|
|
} \
|
|
res |= (static_cast<uint64_t>(wt_op != 0 ? ws_op % wt_op : 0) & mask) \
|
|
<< shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define SRAR_DF(T, lanes, mask) \
|
|
int size_in_bits = kMSARegSize / lanes; \
|
|
for (int i = 0; i < 2; i++) { \
|
|
uint64_t res = 0; \
|
|
for (int j = 0; j < lanes / 2; ++j) { \
|
|
uint64_t shift = size_in_bits * j; \
|
|
T src_op = static_cast<T>((ws[i] >> shift) & mask); \
|
|
T shift_op = ((wt[i] >> shift) & mask) % size_in_bits; \
|
|
uint32_t bit = shift_op == 0 ? 0 : src_op >> (shift_op - 1) & 1; \
|
|
res |= (static_cast<uint64_t>(ArithmeticShiftRight(src_op, shift_op) + \
|
|
bit) & \
|
|
mask) \
|
|
<< shift; \
|
|
} \
|
|
wd[i] = res; \
|
|
}
|
|
|
|
#define PCKEV_DF(T, lanes, mask) \
|
|
T* ws_p = reinterpret_cast<T*>(ws); \
|
|
T* wt_p = reinterpret_cast<T*>(wt); \
|
|
T* wd_p = reinterpret_cast<T*>(wd); \
|
|
for (int i = 0; i < lanes / 2; ++i) { \
|
|
wd_p[i] = wt_p[2 * i]; \
|
|
wd_p[i + lanes / 2] = ws_p[2 * i]; \
|
|
}
|
|
|
|
#define PCKOD_DF(T, lanes, mask) \
|
|
T* ws_p = reinterpret_cast<T*>(ws); \
|
|
T* wt_p = reinterpret_cast<T*>(wt); \
|
|
T* wd_p = reinterpret_cast<T*>(wd); \
|
|
for (int i = 0; i < lanes / 2; ++i) { \
|
|
wd_p[i] = wt_p[2 * i + 1]; \
|
|
wd_p[i + lanes / 2] = ws_p[2 * i + 1]; \
|
|
}
|
|
|
|
#define ILVL_DF(T, lanes, mask) \
|
|
T* ws_p = reinterpret_cast<T*>(ws); \
|
|
T* wt_p = reinterpret_cast<T*>(wt); \
|
|
T* wd_p = reinterpret_cast<T*>(wd); \
|
|
for (int i = 0; i < lanes / 2; ++i) { \
|
|
wd_p[2 * i] = wt_p[i + lanes / 2]; \
|
|
wd_p[2 * i + 1] = ws_p[i + lanes / 2]; \
|
|
}
|
|
|
|
#define ILVR_DF(T, lanes, mask) \
|
|
T* ws_p = reinterpret_cast<T*>(ws); \
|
|
T* wt_p = reinterpret_cast<T*>(wt); \
|
|
T* wd_p = reinterpret_cast<T*>(wd); \
|
|
for (int i = 0; i < lanes / 2; ++i) { \
|
|
wd_p[2 * i] = wt_p[i]; \
|
|
wd_p[2 * i + 1] = ws_p[i]; \
|
|
}
|
|
|
|
#define ILVEV_DF(T, lanes, mask) \
|
|
T* ws_p = reinterpret_cast<T*>(ws); \
|
|
T* wt_p = reinterpret_cast<T*>(wt); \
|
|
T* wd_p = reinterpret_cast<T*>(wd); \
|
|
for (int i = 0; i < lanes / 2; ++i) { \
|
|
wd_p[2 * i] = wt_p[2 * i]; \
|
|
wd_p[2 * i + 1] = ws_p[2 * i]; \
|
|
}
|
|
|
|
#define ILVOD_DF(T, lanes, mask) \
|
|
T* ws_p = reinterpret_cast<T*>(ws); \
|
|
T* wt_p = reinterpret_cast<T*>(wt); \
|
|
T* wd_p = reinterpret_cast<T*>(wd); \
|
|
for (int i = 0; i < lanes / 2; ++i) { \
|
|
wd_p[2 * i] = wt_p[2 * i + 1]; \
|
|
wd_p[2 * i + 1] = ws_p[2 * i + 1]; \
|
|
}
|
|
|
|
#define VSHF_DF(T, lanes, mask) \
|
|
T* ws_p = reinterpret_cast<T*>(ws); \
|
|
T* wt_p = reinterpret_cast<T*>(wt); \
|
|
T* wd_p = reinterpret_cast<T*>(wd); \
|
|
const int mask_not_valid = 0xC0; \
|
|
const int mask_6bits = 0x3F; \
|
|
for (int i = 0; i < lanes; ++i) { \
|
|
if ((wd_p[i] & mask_not_valid)) { \
|
|
wd_p[i] = 0; \
|
|
} else { \
|
|
int k = (wd_p[i] & mask_6bits) % (lanes * 2); \
|
|
wd_p[i] = k > lanes ? ws_p[k - lanes] : wt_p[k]; \
|
|
} \
|
|
}
|
|
|
|
#define HADD_DF(T, T_small, lanes) \
|
|
T_small* ws_p = reinterpret_cast<T_small*>(ws); \
|
|
T_small* wt_p = reinterpret_cast<T_small*>(wt); \
|
|
T* wd_p = reinterpret_cast<T*>(wd); \
|
|
for (int i = 0; i < lanes; ++i) { \
|
|
wd_p[i] = static_cast<T>(ws_p[2 * i + 1]) + static_cast<T>(wt_p[2 * i]); \
|
|
}
|
|
|
|
#define HSUB_DF(T, T_small, lanes) \
|
|
T_small* ws_p = reinterpret_cast<T_small*>(ws); \
|
|
T_small* wt_p = reinterpret_cast<T_small*>(wt); \
|
|
T* wd_p = reinterpret_cast<T*>(wd); \
|
|
for (int i = 0; i < lanes; ++i) { \
|
|
wd_p[i] = static_cast<T>(ws_p[2 * i + 1]) - static_cast<T>(wt_p[2 * i]); \
|
|
}
|
|
|
|
#define TEST_CASE(V) \
|
|
V(sll_b, SLL_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(sll_h, SLL_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(sll_w, SLL_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(sll_d, SLL_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(srl_b, SRL_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(srl_h, SRL_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(srl_w, SRL_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(srl_d, SRL_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(bclr_b, BCRL_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(bclr_h, BCRL_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(bclr_w, BCRL_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(bclr_d, BCRL_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(bset_b, BSET_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(bset_h, BSET_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(bset_w, BSET_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(bset_d, BSET_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(bneg_b, BNEG_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(bneg_h, BNEG_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(bneg_w, BNEG_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(bneg_d, BNEG_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(binsl_b, BINSL_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(binsl_h, BINSL_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(binsl_w, BINSL_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(binsl_d, BINSL_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(binsr_b, BINSR_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(binsr_h, BINSR_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(binsr_w, BINSR_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(binsr_d, BINSR_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(addv_b, ADDV_DF, int8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(addv_h, ADDV_DF, int16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(addv_w, ADDV_DF, int32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(addv_d, ADDV_DF, int64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(subv_b, SUBV_DF, int8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(subv_h, SUBV_DF, int16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(subv_w, SUBV_DF, int32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(subv_d, SUBV_DF, int64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(max_s_b, MAX_DF, int8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(max_s_h, MAX_DF, int16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(max_s_w, MAX_DF, int32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(max_s_d, MAX_DF, int64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(max_u_b, MAX_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(max_u_h, MAX_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(max_u_w, MAX_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(max_u_d, MAX_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(min_s_b, MIN_DF, int8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(min_s_h, MIN_DF, int16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(min_s_w, MIN_DF, int32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(min_s_d, MIN_DF, int64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(min_u_b, MIN_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(min_u_h, MIN_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(min_u_w, MIN_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(min_u_d, MIN_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(max_a_b, MAXA_DF, int8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(max_a_h, MAXA_DF, int16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(max_a_w, MAXA_DF, int32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(max_a_d, MAXA_DF, int64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(min_a_b, MINA_DF, int8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(min_a_h, MINA_DF, int16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(min_a_w, MINA_DF, int32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(min_a_d, MINA_DF, int64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(ceq_b, CEQ_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(ceq_h, CEQ_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(ceq_w, CEQ_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(ceq_d, CEQ_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(clt_s_b, CLT_DF, int8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(clt_s_h, CLT_DF, int16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(clt_s_w, CLT_DF, int32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(clt_s_d, CLT_DF, int64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(clt_u_b, CLT_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(clt_u_h, CLT_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(clt_u_w, CLT_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(clt_u_d, CLT_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(cle_s_b, CLE_DF, int8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(cle_s_h, CLE_DF, int16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(cle_s_w, CLE_DF, int32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(cle_s_d, CLE_DF, int64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(cle_u_b, CLE_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(cle_u_h, CLE_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(cle_u_w, CLE_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(cle_u_d, CLE_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(add_a_b, ADD_A_DF, int8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(add_a_h, ADD_A_DF, int16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(add_a_w, ADD_A_DF, int32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(add_a_d, ADD_A_DF, int64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(adds_a_b, ADDS_A_DF, int8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(adds_a_h, ADDS_A_DF, int16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(adds_a_w, ADDS_A_DF, int32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(adds_a_d, ADDS_A_DF, int64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(adds_s_b, ADDS_DF, int8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(adds_s_h, ADDS_DF, int16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(adds_s_w, ADDS_DF, int32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(adds_s_d, ADDS_DF, int64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(adds_u_b, ADDS_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(adds_u_h, ADDS_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(adds_u_w, ADDS_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(adds_u_d, ADDS_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(ave_s_b, AVE_DF, int8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(ave_s_h, AVE_DF, int16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(ave_s_w, AVE_DF, int32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(ave_s_d, AVE_DF, int64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(ave_u_b, AVE_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(ave_u_h, AVE_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(ave_u_w, AVE_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(ave_u_d, AVE_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(aver_s_b, AVER_DF, int8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(aver_s_h, AVER_DF, int16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(aver_s_w, AVER_DF, int32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(aver_s_d, AVER_DF, int64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(aver_u_b, AVER_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(aver_u_h, AVER_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(aver_u_w, AVER_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(aver_u_d, AVER_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(subs_s_b, SUBS_DF, int8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(subs_s_h, SUBS_DF, int16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(subs_s_w, SUBS_DF, int32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(subs_s_d, SUBS_DF, int64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(subs_u_b, SUBS_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(subs_u_h, SUBS_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(subs_u_w, SUBS_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(subs_u_d, SUBS_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(subsus_u_b, SUBSUS_U_DF, int8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(subsus_u_h, SUBSUS_U_DF, int16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(subsus_u_w, SUBSUS_U_DF, int32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(subsus_u_d, SUBSUS_U_DF, int64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(subsuu_s_b, SUBSUU_S_DF, int8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(subsuu_s_h, SUBSUU_S_DF, int16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(subsuu_s_w, SUBSUU_S_DF, int32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(subsuu_s_d, SUBSUU_S_DF, int64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(asub_s_b, ASUB_S_DF, int8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(asub_s_h, ASUB_S_DF, int16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(asub_s_w, ASUB_S_DF, int32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(asub_s_d, ASUB_S_DF, int64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(asub_u_b, ASUB_U_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(asub_u_h, ASUB_U_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(asub_u_w, ASUB_U_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(asub_u_d, ASUB_U_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(mulv_b, MULV_DF, int8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(mulv_h, MULV_DF, int16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(mulv_w, MULV_DF, int32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(mulv_d, MULV_DF, int64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(maddv_b, MADDV_DF, int8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(maddv_h, MADDV_DF, int16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(maddv_w, MADDV_DF, int32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(maddv_d, MADDV_DF, int64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(msubv_b, MSUBV_DF, int8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(msubv_h, MSUBV_DF, int16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(msubv_w, MSUBV_DF, int32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(msubv_d, MSUBV_DF, int64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(div_s_b, DIV_DF, int8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(div_s_h, DIV_DF, int16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(div_s_w, DIV_DF, int32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(div_s_d, DIV_DF, int64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(div_u_b, DIV_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(div_u_h, DIV_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(div_u_w, DIV_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(div_u_d, DIV_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(mod_s_b, MOD_DF, int8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(mod_s_h, MOD_DF, int16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(mod_s_w, MOD_DF, int32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(mod_s_d, MOD_DF, int64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(mod_u_b, MOD_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(mod_u_h, MOD_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(mod_u_w, MOD_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(mod_u_d, MOD_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(srlr_b, SRAR_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(srlr_h, SRAR_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(srlr_w, SRAR_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(srlr_d, SRAR_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(pckev_b, PCKEV_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(pckev_h, PCKEV_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(pckev_w, PCKEV_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(pckev_d, PCKEV_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(pckod_b, PCKOD_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(pckod_h, PCKOD_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(pckod_w, PCKOD_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(pckod_d, PCKOD_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(ilvl_b, ILVL_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(ilvl_h, ILVL_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(ilvl_w, ILVL_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(ilvl_d, ILVL_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(ilvr_b, ILVR_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(ilvr_h, ILVR_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(ilvr_w, ILVR_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(ilvr_d, ILVR_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(ilvev_b, ILVEV_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(ilvev_h, ILVEV_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(ilvev_w, ILVEV_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(ilvev_d, ILVEV_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(ilvod_b, ILVOD_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(ilvod_h, ILVOD_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(ilvod_w, ILVOD_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(ilvod_d, ILVOD_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(vshf_b, VSHF_DF, uint8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(vshf_h, VSHF_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(vshf_w, VSHF_DF, uint32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(vshf_d, VSHF_DF, uint64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(hadd_s_h, HADD_DF, int16_t, int8_t, kMSALanesHalf) \
|
|
V(hadd_s_w, HADD_DF, int32_t, int16_t, kMSALanesWord) \
|
|
V(hadd_s_d, HADD_DF, int64_t, int32_t, kMSALanesDword) \
|
|
V(hadd_u_h, HADD_DF, uint16_t, uint8_t, kMSALanesHalf) \
|
|
V(hadd_u_w, HADD_DF, uint32_t, uint16_t, kMSALanesWord) \
|
|
V(hadd_u_d, HADD_DF, uint64_t, uint32_t, kMSALanesDword) \
|
|
V(hsub_s_h, HSUB_DF, int16_t, int8_t, kMSALanesHalf) \
|
|
V(hsub_s_w, HSUB_DF, int32_t, int16_t, kMSALanesWord) \
|
|
V(hsub_s_d, HSUB_DF, int64_t, int32_t, kMSALanesDword) \
|
|
V(hsub_u_h, HSUB_DF, uint16_t, uint8_t, kMSALanesHalf) \
|
|
V(hsub_u_w, HSUB_DF, uint32_t, uint16_t, kMSALanesWord) \
|
|
V(hsub_u_d, HSUB_DF, uint64_t, uint32_t, kMSALanesDword)
|
|
|
|
#define RUN_TEST(instr, verify, type, lanes, mask) \
|
|
run_msa_3r(&tc[i], [](MacroAssembler& assm) { __ instr(w2, w1, w0); }, \
|
|
[](uint64_t* ws, uint64_t* wt, uint64_t* wd) { \
|
|
verify(type, lanes, mask); \
|
|
});
|
|
|
|
for (size_t i = 0; i < arraysize(tc); ++i) {
|
|
TEST_CASE(RUN_TEST)
|
|
}
|
|
|
|
#define RUN_TEST2(instr, verify, type, lanes, mask) \
|
|
for (unsigned i = 0; i < arraysize(tc); i++) { \
|
|
for (unsigned j = 0; j < 3; j++) { \
|
|
for (unsigned k = 0; k < lanes; k++) { \
|
|
type* element = reinterpret_cast<type*>(&tc[i]); \
|
|
element[k + j * lanes] &= std::numeric_limits<type>::max(); \
|
|
} \
|
|
} \
|
|
} \
|
|
run_msa_3r(&tc[i], [](MacroAssembler& assm) { __ instr(w2, w1, w0); }, \
|
|
[](uint64_t* ws, uint64_t* wt, uint64_t* wd) { \
|
|
verify(type, lanes, mask); \
|
|
});
|
|
|
|
#define TEST_CASE2(V) \
|
|
V(sra_b, SRA_DF, int8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(sra_h, SRA_DF, int16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(sra_w, SRA_DF, int32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(sra_d, SRA_DF, int64_t, kMSALanesDword, UINT64_MAX) \
|
|
V(srar_b, SRAR_DF, int8_t, kMSALanesByte, UINT8_MAX) \
|
|
V(srar_h, SRAR_DF, int16_t, kMSALanesHalf, UINT16_MAX) \
|
|
V(srar_w, SRAR_DF, int32_t, kMSALanesWord, UINT32_MAX) \
|
|
V(srar_d, SRAR_DF, int64_t, kMSALanesDword, UINT64_MAX)
|
|
|
|
for (size_t i = 0; i < arraysize(tc); ++i) {
|
|
TEST_CASE2(RUN_TEST2)
|
|
}
|
|
|
|
#undef TEST_CASE
|
|
#undef TEST_CASE2
|
|
#undef RUN_TEST
|
|
#undef RUN_TEST2
|
|
#undef SLL_DF
|
|
#undef SRL_DF
|
|
#undef SRA_DF
|
|
#undef BCRL_DF
|
|
#undef BSET_DF
|
|
#undef BNEG_DF
|
|
#undef BINSL_DF
|
|
#undef BINSR_DF
|
|
#undef ADDV_DF
|
|
#undef SUBV_DF
|
|
#undef MAX_DF
|
|
#undef MIN_DF
|
|
#undef MAXA_DF
|
|
#undef MINA_DF
|
|
#undef CEQ_DF
|
|
#undef CLT_DF
|
|
#undef CLE_DF
|
|
#undef ADD_A_DF
|
|
#undef ADDS_A_DF
|
|
#undef ADDS_DF
|
|
#undef AVE_DF
|
|
#undef AVER_DF
|
|
#undef SUBS_DF
|
|
#undef SUBSUS_U_DF
|
|
#undef SUBSUU_S_DF
|
|
#undef ASUB_S_DF
|
|
#undef ASUB_U_DF
|
|
#undef MULV_DF
|
|
#undef MADDV_DF
|
|
#undef MSUBV_DF
|
|
#undef DIV_DF
|
|
#undef MOD_DF
|
|
#undef SRAR_DF
|
|
#undef PCKEV_DF
|
|
#undef PCKOD_DF
|
|
#undef ILVL_DF
|
|
#undef ILVR_DF
|
|
#undef ILVEV_DF
|
|
#undef ILVOD_DF
|
|
#undef VSHF_DF
|
|
#undef HADD_DF
|
|
#undef HSUB_DF
|
|
} // namespace internal
|
|
|
|
struct TestCaseMsa3RF {
|
|
uint64_t ws_lo;
|
|
uint64_t ws_hi;
|
|
uint64_t wt_lo;
|
|
uint64_t wt_hi;
|
|
uint64_t wd_lo;
|
|
uint64_t wd_hi;
|
|
};
|
|
|
|
struct ExpectedResult_MSA3RF {
|
|
uint64_t exp_res_lo;
|
|
uint64_t exp_res_hi;
|
|
};
|
|
|
|
template <typename Func>
|
|
void run_msa_3rf(const struct TestCaseMsa3RF* input,
|
|
const struct ExpectedResult_MSA3RF* output,
|
|
Func Generate2RInstructionFunc) {
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
|
|
MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes);
|
|
CpuFeatureScope fscope(&assm, MIPS_SIMD);
|
|
msa_reg_t res;
|
|
|
|
load_elements_of_vector(
|
|
&assm, reinterpret_cast<const uint64_t*>(&input->ws_lo), w0, t0, t1);
|
|
load_elements_of_vector(
|
|
&assm, reinterpret_cast<const uint64_t*>(&input->wt_lo), w1, t0, t1);
|
|
load_elements_of_vector(
|
|
&assm, reinterpret_cast<const uint64_t*>(&input->wd_lo), w2, t0, t1);
|
|
Generate2RInstructionFunc(assm);
|
|
store_elements_of_vector(&assm, w2, a0);
|
|
|
|
__ jr(ra);
|
|
__ nop();
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::STUB).Build();
|
|
#ifdef OBJECT_PRINT
|
|
code->Print(std::cout);
|
|
#endif
|
|
auto f = GeneratedCode<F3>::FromCode(*code);
|
|
|
|
(f.Call(&res, 0, 0, 0, 0));
|
|
|
|
CHECK_EQ(output->exp_res_lo, res.d[0]);
|
|
CHECK_EQ(output->exp_res_hi, res.d[1]);
|
|
}
|
|
|
|
struct TestCaseMsa3RF_F {
|
|
float ws_1, ws_2, ws_3, ws_4;
|
|
float wt_1, wt_2, wt_3, wt_4;
|
|
float wd_1, wd_2, wd_3, wd_4;
|
|
};
|
|
struct ExpRes_32I {
|
|
int32_t exp_res_1;
|
|
int32_t exp_res_2;
|
|
int32_t exp_res_3;
|
|
int32_t exp_res_4;
|
|
};
|
|
|
|
struct TestCaseMsa3RF_D {
|
|
double ws_lo, ws_hi;
|
|
double wt_lo, wt_hi;
|
|
double wd_lo, wd_hi;
|
|
};
|
|
struct ExpRes_64I {
|
|
int64_t exp_res_lo;
|
|
int64_t exp_res_hi;
|
|
};
|
|
|
|
TEST(MSA_floating_point_quiet_compare) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
const float qnan_f = std::numeric_limits<float>::quiet_NaN();
|
|
const double qnan_d = std::numeric_limits<double>::quiet_NaN();
|
|
const float inf_f = std::numeric_limits<float>::infinity();
|
|
const double inf_d = std::numeric_limits<double>::infinity();
|
|
const int32_t ones = -1;
|
|
|
|
const struct TestCaseMsa3RF_F tc_w[]{
|
|
{qnan_f, -qnan_f, inf_f, 2.14e9f, // ws
|
|
qnan_f, 0.f, qnan_f, -2.14e9f, // wt
|
|
0, 0, 0, 0}, // wd
|
|
{inf_f, -inf_f, -3.4e38f, 1.5e-45f, -inf_f, -inf_f, -inf_f, inf_f, 0, 0,
|
|
0, 0},
|
|
{0.f, 19.871e24f, -1.5e-45f, -1.5e-45f, -19.871e24f, 19.871e24f, 1.5e-45f,
|
|
-1.5e-45f, 0, 0, 0, 0}};
|
|
|
|
const struct TestCaseMsa3RF_D tc_d[]{
|
|
// ws_lo, ws_hi, wt_lo, wt_hi, wd_lo, wd_hi
|
|
{qnan_d, -qnan_d, qnan_f, 0., 0, 0},
|
|
{inf_d, 9.22e18, qnan_d, -9.22e18, 0, 0},
|
|
{inf_d, inf_d, -inf_d, inf_d, 0, 0},
|
|
{-2.3e-308, 5e-324, -inf_d, inf_d, 0, 0},
|
|
{0., 24.1e87, -1.6e308, 24.1e87, 0, 0},
|
|
{-5e-324, -5e-324, 5e-324, -5e-324, 0, 0}};
|
|
|
|
const struct ExpectedResult_MSA3RF exp_res_fcaf = {0, 0};
|
|
const struct ExpRes_32I exp_res_fcun_w[] = {
|
|
{ones, ones, ones, 0}, {0, 0, 0, 0}, {0, 0, 0, 0}};
|
|
const struct ExpRes_64I exp_res_fcun_d[] = {{ones, ones}, {ones, 0}, {0, 0},
|
|
{0, 0}, {0, 0}, {0, 0}};
|
|
const struct ExpRes_32I exp_res_fceq_w[] = {
|
|
{0, 0, 0, 0}, {0, ones, 0, 0}, {0, ones, 0, ones}};
|
|
const struct ExpRes_64I exp_res_fceq_d[] = {{0, 0}, {0, 0}, {0, ones},
|
|
{0, 0}, {0, ones}, {0, ones}};
|
|
const struct ExpRes_32I exp_res_fcueq_w[] = {
|
|
{ones, ones, ones, 0}, {0, ones, 0, 0}, {0, ones, 0, ones}};
|
|
const struct ExpRes_64I exp_res_fcueq_d[] = {
|
|
{ones, ones}, {ones, 0}, {0, ones}, {0, 0}, {0, ones}, {0, ones}};
|
|
const struct ExpRes_32I exp_res_fclt_w[] = {
|
|
{0, 0, 0, 0}, {0, 0, 0, ones}, {0, 0, ones, 0}};
|
|
const struct ExpRes_64I exp_res_fclt_d[] = {{0, 0}, {0, 0}, {0, 0},
|
|
{0, ones}, {0, 0}, {ones, 0}};
|
|
const struct ExpRes_32I exp_res_fcult_w[] = {
|
|
{ones, ones, ones, 0}, {0, 0, 0, ones}, {0, 0, ones, 0}};
|
|
const struct ExpRes_64I exp_res_fcult_d[] = {
|
|
{ones, ones}, {ones, 0}, {0, 0}, {0, ones}, {0, 0}, {ones, 0}};
|
|
const struct ExpRes_32I exp_res_fcle_w[] = {
|
|
{0, 0, 0, 0}, {0, ones, 0, ones}, {0, ones, ones, ones}};
|
|
const struct ExpRes_64I exp_res_fcle_d[] = {
|
|
{0, 0}, {0, 0}, {0, ones}, {0, ones}, {0, ones}, {ones, ones}};
|
|
const struct ExpRes_32I exp_res_fcule_w[] = {
|
|
{ones, ones, ones, 0}, {0, ones, 0, ones}, {0, ones, ones, ones}};
|
|
const struct ExpRes_64I exp_res_fcule_d[] = {
|
|
{ones, ones}, {ones, 0}, {0, ones}, {0, ones}, {0, ones}, {ones, ones}};
|
|
const struct ExpRes_32I exp_res_fcor_w[] = {
|
|
{0, 0, 0, ones}, {ones, ones, ones, ones}, {ones, ones, ones, ones}};
|
|
const struct ExpRes_64I exp_res_fcor_d[] = {{0, 0}, {0, ones},
|
|
{ones, ones}, {ones, ones},
|
|
{ones, ones}, {ones, ones}};
|
|
const struct ExpRes_32I exp_res_fcune_w[] = {
|
|
{ones, ones, ones, ones}, {ones, 0, ones, ones}, {ones, 0, ones, 0}};
|
|
const struct ExpRes_64I exp_res_fcune_d[] = {{ones, ones}, {ones, ones},
|
|
{ones, 0}, {ones, ones},
|
|
{ones, 0}, {ones, 0}};
|
|
const struct ExpRes_32I exp_res_fcne_w[] = {
|
|
{0, 0, 0, ones}, {ones, 0, ones, ones}, {ones, 0, ones, 0}};
|
|
const struct ExpRes_64I exp_res_fcne_d[] = {
|
|
{0, 0}, {0, ones}, {ones, 0}, {ones, ones}, {ones, 0}, {ones, 0}};
|
|
|
|
#define TEST_FP_QUIET_COMPARE_W(instruction, src, exp_res) \
|
|
run_msa_3rf(reinterpret_cast<const struct TestCaseMsa3RF*>(src), \
|
|
reinterpret_cast<const struct ExpectedResult_MSA3RF*>(exp_res), \
|
|
[](MacroAssembler& assm) { __ instruction(w2, w0, w1); });
|
|
#define TEST_FP_QUIET_COMPARE_D(instruction, src, exp_res) \
|
|
run_msa_3rf(reinterpret_cast<const struct TestCaseMsa3RF*>(src), \
|
|
reinterpret_cast<const struct ExpectedResult_MSA3RF*>(exp_res), \
|
|
[](MacroAssembler& assm) { __ instruction(w2, w0, w1); });
|
|
|
|
for (uint64_t i = 0; i < arraysize(tc_w); i++) {
|
|
TEST_FP_QUIET_COMPARE_W(fcaf_w, &tc_w[i], &exp_res_fcaf)
|
|
TEST_FP_QUIET_COMPARE_W(fcun_w, &tc_w[i], &exp_res_fcun_w[i])
|
|
TEST_FP_QUIET_COMPARE_W(fceq_w, &tc_w[i], &exp_res_fceq_w[i])
|
|
TEST_FP_QUIET_COMPARE_W(fcueq_w, &tc_w[i], &exp_res_fcueq_w[i])
|
|
TEST_FP_QUIET_COMPARE_W(fclt_w, &tc_w[i], &exp_res_fclt_w[i])
|
|
TEST_FP_QUIET_COMPARE_W(fcult_w, &tc_w[i], &exp_res_fcult_w[i])
|
|
TEST_FP_QUIET_COMPARE_W(fcle_w, &tc_w[i], &exp_res_fcle_w[i])
|
|
TEST_FP_QUIET_COMPARE_W(fcule_w, &tc_w[i], &exp_res_fcule_w[i])
|
|
TEST_FP_QUIET_COMPARE_W(fcor_w, &tc_w[i], &exp_res_fcor_w[i])
|
|
TEST_FP_QUIET_COMPARE_W(fcune_w, &tc_w[i], &exp_res_fcune_w[i])
|
|
TEST_FP_QUIET_COMPARE_W(fcne_w, &tc_w[i], &exp_res_fcne_w[i])
|
|
}
|
|
for (uint64_t i = 0; i < arraysize(tc_d); i++) {
|
|
TEST_FP_QUIET_COMPARE_D(fcaf_d, &tc_d[i], &exp_res_fcaf)
|
|
TEST_FP_QUIET_COMPARE_D(fcun_d, &tc_d[i], &exp_res_fcun_d[i])
|
|
TEST_FP_QUIET_COMPARE_D(fceq_d, &tc_d[i], &exp_res_fceq_d[i])
|
|
TEST_FP_QUIET_COMPARE_D(fcueq_d, &tc_d[i], &exp_res_fcueq_d[i])
|
|
TEST_FP_QUIET_COMPARE_D(fclt_d, &tc_d[i], &exp_res_fclt_d[i])
|
|
TEST_FP_QUIET_COMPARE_D(fcult_d, &tc_d[i], &exp_res_fcult_d[i])
|
|
TEST_FP_QUIET_COMPARE_D(fcle_d, &tc_d[i], &exp_res_fcle_d[i])
|
|
TEST_FP_QUIET_COMPARE_D(fcule_d, &tc_d[i], &exp_res_fcule_d[i])
|
|
TEST_FP_QUIET_COMPARE_D(fcor_d, &tc_d[i], &exp_res_fcor_d[i])
|
|
TEST_FP_QUIET_COMPARE_D(fcune_d, &tc_d[i], &exp_res_fcune_d[i])
|
|
TEST_FP_QUIET_COMPARE_D(fcne_d, &tc_d[i], &exp_res_fcne_d[i])
|
|
}
|
|
#undef TEST_FP_QUIET_COMPARE_W
|
|
#undef TEST_FP_QUIET_COMPARE_D
|
|
}
|
|
|
|
template <typename T>
|
|
inline const T* fadd_function(const T* src1, const T* src2, const T* src3,
|
|
T* dst) {
|
|
for (uint64_t i = 0; i < kMSALanesByte / sizeof(T); i++) {
|
|
dst[i] = src1[i] + src2[i];
|
|
}
|
|
return dst;
|
|
}
|
|
template <typename T>
|
|
inline const T* fsub_function(const T* src1, const T* src2, const T* src3,
|
|
T* dst) {
|
|
for (uint64_t i = 0; i < kMSALanesByte / sizeof(T); i++) {
|
|
dst[i] = src1[i] - src2[i];
|
|
}
|
|
return dst;
|
|
}
|
|
template <typename T>
|
|
inline const T* fmul_function(const T* src1, const T* src2, const T* src3,
|
|
T* dst) {
|
|
for (uint64_t i = 0; i < kMSALanesByte / sizeof(T); i++) {
|
|
dst[i] = src1[i] * src2[i];
|
|
}
|
|
return dst;
|
|
}
|
|
template <typename T>
|
|
inline const T* fdiv_function(const T* src1, const T* src2, const T* src3,
|
|
T* dst) {
|
|
for (uint64_t i = 0; i < kMSALanesByte / sizeof(T); i++) {
|
|
dst[i] = src1[i] / src2[i];
|
|
}
|
|
return dst;
|
|
}
|
|
template <typename T>
|
|
inline const T* fmadd_function(const T* src1, const T* src2, const T* src3,
|
|
T* dst) {
|
|
for (uint64_t i = 0; i < kMSALanesByte / sizeof(T); i++) {
|
|
dst[i] = std::fma(src1[i], src2[i], src3[i]);
|
|
}
|
|
return dst;
|
|
}
|
|
template <typename T>
|
|
inline const T* fmsub_function(const T* src1, const T* src2, const T* src3,
|
|
T* dst) {
|
|
for (uint64_t i = 0; i < kMSALanesByte / sizeof(T); i++) {
|
|
dst[i] = std::fma(src1[i], -src2[i], src3[i]);
|
|
}
|
|
return dst;
|
|
}
|
|
|
|
TEST(MSA_floating_point_arithmetic) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
const float inf_f = std::numeric_limits<float>::infinity();
|
|
const double inf_d = std::numeric_limits<double>::infinity();
|
|
|
|
const struct TestCaseMsa3RF_F tc_w[] = {
|
|
{0.3, -2.14e13f, inf_f, 0.f, // ws
|
|
-inf_f, std::sqrt(8.e-26f), -23.e34, -2.14e9f, // wt
|
|
-1e30f, 4.6e12f, 0, 2.14e9f}, // wd
|
|
{3.4e38f, -1.2e-38f, 1e19f, -1e19f, 3.4e38f, 1.2e-38f, -1e19f, -1e-19f,
|
|
3.4e38f, 1.2e-38f * 3, 3.4e38f, -4e19f},
|
|
{-3e-31f, 3e10f, 1e25f, 123.f, 1e-14f, 1e-34f, 4e25f, 321.f, 3e-17f,
|
|
2e-24f, 2.f, -123456.f}};
|
|
|
|
const struct TestCaseMsa3RF_D tc_d[] = {
|
|
// ws_lo, ws_hi, wt_lo, wt_hi, wd_lo, wd_hi
|
|
{0.3, -2.14e103, -inf_d, std::sqrt(8.e-206), -1e30, 4.6e102},
|
|
{inf_d, 0., -23.e304, -2.104e9, 0, 2.104e9},
|
|
{3.4e307, -1.2e-307, 3.4e307, 1.2e-307, 3.4e307, 1.2e-307 * 3},
|
|
{1e154, -1e154, -1e154, -1e-154, 2.9e38, -4e19},
|
|
{-3e-301, 3e100, 1e-104, 1e-304, 3e-107, 2e-204},
|
|
{1e205, 123., 4e205, 321., 2., -123456.}};
|
|
|
|
struct ExpectedResult_MSA3RF dst_container;
|
|
|
|
#define FP_ARITHMETIC_DF_W(instr, function, src1, src2, src3) \
|
|
run_msa_3rf( \
|
|
reinterpret_cast<const struct TestCaseMsa3RF*>(src1), \
|
|
reinterpret_cast<const struct ExpectedResult_MSA3RF*>(function( \
|
|
src1, src2, src3, reinterpret_cast<float*>(&dst_container))), \
|
|
[](MacroAssembler& assm) { __ instr(w2, w0, w1); });
|
|
|
|
#define FP_ARITHMETIC_DF_D(instr, function, src1, src2, src3) \
|
|
run_msa_3rf( \
|
|
reinterpret_cast<const struct TestCaseMsa3RF*>(src1), \
|
|
reinterpret_cast<const struct ExpectedResult_MSA3RF*>(function( \
|
|
src1, src2, src3, reinterpret_cast<double*>(&dst_container))), \
|
|
[](MacroAssembler& assm) { __ instr(w2, w0, w1); });
|
|
|
|
for (uint64_t i = 0; i < arraysize(tc_w); i++) {
|
|
FP_ARITHMETIC_DF_W(fadd_w, fadd_function, &tc_w[i].ws_1, &tc_w[i].wt_1,
|
|
&tc_w[i].wd_1)
|
|
FP_ARITHMETIC_DF_W(fsub_w, fsub_function, &tc_w[i].ws_1, &tc_w[i].wt_1,
|
|
&tc_w[i].wd_1)
|
|
FP_ARITHMETIC_DF_W(fmul_w, fmul_function, &tc_w[i].ws_1, &tc_w[i].wt_1,
|
|
&tc_w[i].wd_1)
|
|
FP_ARITHMETIC_DF_W(fdiv_w, fdiv_function, &tc_w[i].ws_1, &tc_w[i].wt_1,
|
|
&tc_w[i].wd_1)
|
|
FP_ARITHMETIC_DF_W(fmadd_w, fmadd_function, &tc_w[i].ws_1, &tc_w[i].wt_1,
|
|
&tc_w[i].wd_1)
|
|
FP_ARITHMETIC_DF_W(fmsub_w, fmsub_function, &tc_w[i].ws_1, &tc_w[i].wt_1,
|
|
&tc_w[i].wd_1)
|
|
}
|
|
for (uint64_t i = 0; i < arraysize(tc_d); i++) {
|
|
FP_ARITHMETIC_DF_D(fadd_d, fadd_function, &tc_d[i].ws_lo, &tc_d[i].wt_lo,
|
|
&tc_d[i].wd_lo)
|
|
FP_ARITHMETIC_DF_D(fsub_d, fsub_function, &tc_d[i].ws_lo, &tc_d[i].wt_lo,
|
|
&tc_d[i].wd_lo)
|
|
FP_ARITHMETIC_DF_D(fmul_d, fmul_function, &tc_d[i].ws_lo, &tc_d[i].wt_lo,
|
|
&tc_d[i].wd_lo)
|
|
FP_ARITHMETIC_DF_D(fdiv_d, fdiv_function, &tc_d[i].ws_lo, &tc_d[i].wt_lo,
|
|
&tc_d[i].wd_lo)
|
|
FP_ARITHMETIC_DF_D(fmadd_d, fmadd_function, &tc_d[i].ws_lo, &tc_d[i].wt_lo,
|
|
&tc_d[i].wd_lo)
|
|
FP_ARITHMETIC_DF_D(fmsub_d, fmsub_function, &tc_d[i].ws_lo, &tc_d[i].wt_lo,
|
|
&tc_d[i].wd_lo)
|
|
}
|
|
#undef FP_ARITHMETIC_DF_W
|
|
#undef FP_ARITHMETIC_DF_D
|
|
}
|
|
|
|
struct ExpRes_F {
|
|
float exp_res_1;
|
|
float exp_res_2;
|
|
float exp_res_3;
|
|
float exp_res_4;
|
|
};
|
|
|
|
struct ExpRes_D {
|
|
double exp_res_1;
|
|
double exp_res_2;
|
|
};
|
|
|
|
TEST(MSA_fmin_fmin_a_fmax_fmax_a) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
const float inf_f = std::numeric_limits<float>::infinity();
|
|
const double inf_d = std::numeric_limits<double>::infinity();
|
|
|
|
const struct TestCaseMsa3RF_F tc_w[] = {
|
|
{0.3f, -2.14e13f, inf_f, -0.f, // ws
|
|
-inf_f, -std::sqrt(8.e26f), -23.e34f, -2.14e9f, // wt
|
|
0, 0, 0, 0}, // wd
|
|
{3.4e38f, 1.2e-41f, 1e19f, 1e19f, // ws
|
|
3.4e38f, -1.1e-41f, -1e-42f, -1e29f, // wt
|
|
0, 0, 0, 0}}; // wd
|
|
|
|
const struct TestCaseMsa3RF_D tc_d[] = {
|
|
// ws_lo, ws_hi, wt_lo, wt_hi, wd_lo, wd_hi
|
|
{0.3, -2.14e103, -inf_d, -std::sqrt(8e206), 0, 0},
|
|
{inf_d, -0., -23e304, -2.14e90, 0, 0},
|
|
{3.4e307, 1.2e-320, 3.4e307, -1.1e-320, 0, 0},
|
|
{1e154, 1e154, -1e-321, -1e174, 0, 0}};
|
|
|
|
const struct ExpRes_F exp_res_fmax_w[] = {{0.3f, -2.14e13f, inf_f, -0.f},
|
|
{3.4e38f, 1.2e-41f, 1e19f, 1e19f}};
|
|
const struct ExpRes_F exp_res_fmax_a_w[] = {
|
|
{-inf_f, -std::sqrt(8e26f), inf_f, -2.14e9f},
|
|
{3.4e38f, 1.2e-41f, 1e19f, -1e29f}};
|
|
const struct ExpRes_F exp_res_fmin_w[] = {
|
|
{-inf_f, -std::sqrt(8.e26f), -23e34f, -2.14e9f},
|
|
{3.4e38f, -1.1e-41f, -1e-42f, -1e29f}};
|
|
const struct ExpRes_F exp_res_fmin_a_w[] = {
|
|
{0.3, -2.14e13f, -23.e34f, -0.f}, {3.4e38f, -1.1e-41f, -1e-42f, 1e19f}};
|
|
|
|
const struct ExpRes_D exp_res_fmax_d[] = {
|
|
{0.3, -2.14e103}, {inf_d, -0.}, {3.4e307, 1.2e-320}, {1e154, 1e154}};
|
|
const struct ExpRes_D exp_res_fmax_a_d[] = {{-inf_d, -std::sqrt(8e206)},
|
|
{inf_d, -2.14e90},
|
|
{3.4e307, 1.2e-320},
|
|
{1e154, -1e174}};
|
|
const struct ExpRes_D exp_res_fmin_d[] = {{-inf_d, -std::sqrt(8e206)},
|
|
{-23e304, -2.14e90},
|
|
{3.4e307, -1.1e-320},
|
|
{-1e-321, -1e174}};
|
|
const struct ExpRes_D exp_res_fmin_a_d[] = {
|
|
{0.3, -2.14e103}, {-23e304, -0.}, {3.4e307, -1.1e-320}, {-1e-321, 1e154}};
|
|
|
|
#define TEST_FP_MIN_MAX_W(instruction, src, exp_res) \
|
|
run_msa_3rf(reinterpret_cast<const struct TestCaseMsa3RF*>(src), \
|
|
reinterpret_cast<const struct ExpectedResult_MSA3RF*>(exp_res), \
|
|
[](MacroAssembler& assm) { __ instruction(w2, w0, w1); });
|
|
|
|
#define TEST_FP_MIN_MAX_D(instruction, src, exp_res) \
|
|
run_msa_3rf(reinterpret_cast<const struct TestCaseMsa3RF*>(src), \
|
|
reinterpret_cast<const struct ExpectedResult_MSA3RF*>(exp_res), \
|
|
[](MacroAssembler& assm) { __ instruction(w2, w0, w1); });
|
|
|
|
for (uint64_t i = 0; i < arraysize(tc_w); i++) {
|
|
TEST_FP_MIN_MAX_W(fmax_w, &tc_w[i], &exp_res_fmax_w[i])
|
|
TEST_FP_MIN_MAX_W(fmax_a_w, &tc_w[i], &exp_res_fmax_a_w[i])
|
|
TEST_FP_MIN_MAX_W(fmin_w, &tc_w[i], &exp_res_fmin_w[i])
|
|
TEST_FP_MIN_MAX_W(fmin_a_w, &tc_w[i], &exp_res_fmin_a_w[i])
|
|
}
|
|
|
|
for (uint64_t i = 0; i < arraysize(tc_d); i++) {
|
|
TEST_FP_MIN_MAX_D(fmax_d, &tc_d[i], &exp_res_fmax_d[i])
|
|
TEST_FP_MIN_MAX_D(fmax_a_d, &tc_d[i], &exp_res_fmax_a_d[i])
|
|
TEST_FP_MIN_MAX_D(fmin_d, &tc_d[i], &exp_res_fmin_d[i])
|
|
TEST_FP_MIN_MAX_D(fmin_a_d, &tc_d[i], &exp_res_fmin_a_d[i])
|
|
}
|
|
#undef TEST_FP_MIN_MAX_W
|
|
#undef TEST_FP_MIN_MAX_D
|
|
}
|
|
|
|
struct TestCaseMsa3RF_16I {
|
|
int16_t ws_1, ws_2, ws_3, ws_4, ws_5, ws_6, ws_7, ws_8;
|
|
int16_t wt_1, wt_2, wt_3, wt_4, wt_5, wt_6, wt_7, wt_8;
|
|
int16_t wd_1, wd_2, wd_3, wd_4, wd_5, wd_6, wd_7, wd_8;
|
|
};
|
|
struct ExpRes_16I {
|
|
int16_t exp_res_1;
|
|
int16_t exp_res_2;
|
|
int16_t exp_res_3;
|
|
int16_t exp_res_4;
|
|
int16_t exp_res_5;
|
|
int16_t exp_res_6;
|
|
int16_t exp_res_7;
|
|
int16_t exp_res_8;
|
|
};
|
|
|
|
struct TestCaseMsa3RF_32I {
|
|
int32_t ws_1, ws_2, ws_3, ws_4;
|
|
int32_t wt_1, wt_2, wt_3, wt_4;
|
|
int32_t wd_1, wd_2, wd_3, wd_4;
|
|
};
|
|
|
|
TEST(MSA_fixed_point_arithmetic) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
const struct TestCaseMsa3RF tc_h[]{
|
|
{0x800080007FFF7FFF, 0xE1ED8000FAD3863A, 0x80007FFF00AF7FFF,
|
|
0x800015A77FFFA0EB, 0x7FFF800080007FFF, 0x80007FFF1F207364},
|
|
{0x800080007FFF006A, 0x002AFFC4329AD87B, 0x80007FFF7FFF00F3,
|
|
0xFFECFFB4D0D7F429, 0x80007FFF80007C33, 0x54AC6BBCE53B8C91}};
|
|
|
|
const struct TestCaseMsa3RF tc_w[]{
|
|
{0x8000000080000000, 0x7FFFFFFF7FFFFFFF, 0x800000007FFFFFFF,
|
|
0x00001FF37FFFFFFF, 0x7FFFFFFF80000000, 0x800000007FFFFFFF},
|
|
{0xE1ED035580000000, 0xFAD3863AED462C0B, 0x8000000015A70AEC,
|
|
0x7FFFFFFFA0EBD354, 0x800000007FFFFFFF, 0xD0D7F4291F207364},
|
|
{0x8000000080000000, 0x7FFFFFFF0000DA1F, 0x800000007FFFFFFF,
|
|
0x7FFFFFFF00F39C3B, 0x800000007FFFFFFF, 0x800000007C33F2FD},
|
|
{0x0000AC33FFFF329A, 0x54AC6BBCE53BD87B, 0xFFFFE2B4D0D7F429,
|
|
0x0355ED462C0B1FF3, 0xB5DEB625939DD3F9, 0xE642ADFA69519596}};
|
|
|
|
const struct ExpectedResult_MSA3RF exp_res_mul_q_h[] = {
|
|
{0x7FFF800100AE7FFE, 0x1E13EA59FAD35A74},
|
|
{0x7FFF80017FFE0000, 0xFFFF0000ED5B03A7}};
|
|
const struct ExpectedResult_MSA3RF exp_res_madd_q_h[] = {
|
|
{0x7FFF800080AE7FFF, 0x9E136A5819F37FFF},
|
|
{0x00000000FFFE7C33, 0x54AB6BBCD2969038}};
|
|
const struct ExpectedResult_MSA3RF exp_res_msub_q_h[] = {
|
|
{0xFFFFFFFF80000000, 0x80007FFF244C18EF},
|
|
{0x80007FFF80007C32, 0x54AC6BBBF7DF88E9}};
|
|
const struct ExpectedResult_MSA3RF exp_res_mulr_q_h[] = {
|
|
{0x7FFF800100AF7FFE, 0x1E13EA59FAD35A75},
|
|
{0x7FFF80017FFE0001, 0x00000000ED5B03A8}};
|
|
const struct ExpectedResult_MSA3RF exp_res_maddr_q_h[] = {
|
|
{0x7FFF800080AF7FFF, 0x9E136A5819F37FFF},
|
|
{0x00000000FFFE7C34, 0x54AC6BBCD2969039}};
|
|
const struct ExpectedResult_MSA3RF exp_res_msubr_q_h[] = {
|
|
{0xFFFFFFFF80000001, 0x80007FFF244D18EF},
|
|
{0x80007FFF80007C32, 0x54AC6BBCF7E088E9}};
|
|
|
|
const struct ExpectedResult_MSA3RF exp_res_mul_q_w[] = {
|
|
{0x7FFFFFFF80000001, 0x00001FF27FFFFFFE},
|
|
{0x1E12FCABEA58F514, 0xFAD3863A0DE8DEE1},
|
|
{0x7FFFFFFF80000001, 0x7FFFFFFE0000019F},
|
|
{0xFFFFFFFF00004BAB, 0x0234E1FBF6CA3EE0}};
|
|
const struct ExpectedResult_MSA3RF exp_res_madd_q_w[] = {
|
|
{0x7FFFFFFF80000000, 0x80001FF27FFFFFFF},
|
|
{0x9E12FCAB6A58F513, 0xCBAB7A632D095245},
|
|
{0x0000000000000000, 0xFFFFFFFE7C33F49C},
|
|
{0xB5DEB624939E1FA4, 0xE8778FF5601BD476}};
|
|
const struct ExpectedResult_MSA3RF exp_res_msub_q_w[] = {
|
|
{0xFFFFFFFFFFFFFFFF, 0x8000000000000000},
|
|
{0x800000007FFFFFFF, 0xD6046DEE11379482},
|
|
{0x800000007FFFFFFF, 0x800000007C33F15D},
|
|
{0xB5DEB625939D884D, 0xE40DCBFE728756B5}};
|
|
const struct ExpectedResult_MSA3RF exp_res_mulr_q_w[] = {
|
|
{0x7FFFFFFF80000001, 0x00001FF37FFFFFFE},
|
|
{0x1E12FCABEA58F514, 0xFAD3863A0DE8DEE2},
|
|
{0x7FFFFFFF80000001, 0x7FFFFFFE0000019F},
|
|
{0x0000000000004BAC, 0x0234E1FCF6CA3EE1}};
|
|
const struct ExpectedResult_MSA3RF exp_res_maddr_q_w[] = {
|
|
{0x7FFFFFFF80000000, 0x80001FF37FFFFFFF},
|
|
{0x9E12FCAB6A58F513, 0xCBAB7A632D095246},
|
|
{0x0000000000000000, 0xFFFFFFFE7C33F49C},
|
|
{0xB5DEB625939E1FA5, 0xE8778FF6601BD477}};
|
|
const struct ExpectedResult_MSA3RF exp_res_msubr_q_w[] = {
|
|
{0xFFFFFFFFFFFFFFFF, 0x8000000000000001},
|
|
{0x800000007FFFFFFF, 0xD6046DEF11379482},
|
|
{0x800000007FFFFFFF, 0x800000007C33F15E},
|
|
{0xB5DEB625939D884D, 0xE40DCBFE728756B5}};
|
|
|
|
#define TEST_FIXED_POINT_DF_H(instruction, src, exp_res) \
|
|
run_msa_3rf((src), (exp_res), \
|
|
[](MacroAssembler& assm) { __ instruction(w2, w0, w1); });
|
|
|
|
#define TEST_FIXED_POINT_DF_W(instruction, src, exp_res) \
|
|
run_msa_3rf((src), (exp_res), \
|
|
[](MacroAssembler& assm) { __ instruction(w2, w0, w1); });
|
|
|
|
for (uint64_t i = 0; i < arraysize(tc_h); i++) {
|
|
TEST_FIXED_POINT_DF_H(mul_q_h, &tc_h[i], &exp_res_mul_q_h[i])
|
|
TEST_FIXED_POINT_DF_H(madd_q_h, &tc_h[i], &exp_res_madd_q_h[i])
|
|
TEST_FIXED_POINT_DF_H(msub_q_h, &tc_h[i], &exp_res_msub_q_h[i])
|
|
TEST_FIXED_POINT_DF_H(mulr_q_h, &tc_h[i], &exp_res_mulr_q_h[i])
|
|
TEST_FIXED_POINT_DF_H(maddr_q_h, &tc_h[i], &exp_res_maddr_q_h[i])
|
|
TEST_FIXED_POINT_DF_H(msubr_q_h, &tc_h[i], &exp_res_msubr_q_h[i])
|
|
}
|
|
|
|
for (uint64_t i = 0; i < arraysize(tc_w); i++) {
|
|
TEST_FIXED_POINT_DF_W(mul_q_w, &tc_w[i], &exp_res_mul_q_w[i])
|
|
TEST_FIXED_POINT_DF_W(madd_q_w, &tc_w[i], &exp_res_madd_q_w[i])
|
|
TEST_FIXED_POINT_DF_W(msub_q_w, &tc_w[i], &exp_res_msub_q_w[i])
|
|
TEST_FIXED_POINT_DF_W(mulr_q_w, &tc_w[i], &exp_res_mulr_q_w[i])
|
|
TEST_FIXED_POINT_DF_W(maddr_q_w, &tc_w[i], &exp_res_maddr_q_w[i])
|
|
TEST_FIXED_POINT_DF_W(msubr_q_w, &tc_w[i], &exp_res_msubr_q_w[i])
|
|
}
|
|
#undef TEST_FIXED_POINT_DF_H
|
|
#undef TEST_FIXED_POINT_DF_W
|
|
}
|
|
|
|
TEST(MSA_fexdo) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
const float inf_float = std::numeric_limits<float>::infinity();
|
|
const float nan_float = std::numeric_limits<float>::quiet_NaN();
|
|
const double inf_double = std::numeric_limits<double>::infinity();
|
|
|
|
const struct TestCaseMsa3RF_F tc_w[] = {
|
|
// ws_1, ws_2, ws_3, ws_4, wt_1, wt_2, wt_3, wt_4, wd_1, wd_2, wd_3, wd_4
|
|
{inf_float, nan_float, 66505.f, 65504.f, 6.2e-5f, 5e-5f, -32.42f,
|
|
-inf_float, 0, 0, 0, 0},
|
|
{-0.f, 0.f, 123.567f, -765.321f, -6e-8f, 5.9e-8f, 1e-7f, -1e-20f, 0, 0, 0,
|
|
0},
|
|
{1e-36f, 1e20f, -1e20f, 2e-20f, 6e-8f, -2.9e-8f, -66505.f, -65504.f, 0, 0,
|
|
0, 0}};
|
|
|
|
const struct TestCaseMsa3RF_D tc_d[] = {
|
|
// ws_lo, ws_hi, wt_lo, wt_hi, wd_lo, wd_hi
|
|
{inf_double, -1234., 4e38, 3.4e38, 0, 0},
|
|
{1.2e-38, 1.1e-39, -38.92f, -inf_double, 0, 0},
|
|
{-0., 0., 123.567e31, -765.321e33, 0, 0},
|
|
{-1.5e-45, 1.3e-45, 1e-42, -1e-200, 0, 0},
|
|
{1e-202, 1e158, -1e159, 1e14, 0, 0},
|
|
{1.5e-42, 1.3e-46, -123.567e31, 765.321e33, 0, 0}};
|
|
|
|
const struct ExpRes_16I exp_res_fexdo_w[] = {
|
|
{static_cast<int16_t>(0x0410), static_cast<int16_t>(0x0347),
|
|
static_cast<int16_t>(0xD00D), static_cast<int16_t>(0xFC00),
|
|
static_cast<int16_t>(0x7C00), static_cast<int16_t>(0x7DFF),
|
|
static_cast<int16_t>(0x7C00), static_cast<int16_t>(0x7BFF)},
|
|
{static_cast<int16_t>(0x8001), static_cast<int16_t>(0x0001),
|
|
static_cast<int16_t>(0x0002), static_cast<int16_t>(0x8000),
|
|
static_cast<int16_t>(0x8000), static_cast<int16_t>(0x0000),
|
|
static_cast<int16_t>(0x57B9), static_cast<int16_t>(0xE1FB)},
|
|
{static_cast<int16_t>(0x0001), static_cast<int16_t>(0x8000),
|
|
static_cast<int16_t>(0xFC00), static_cast<int16_t>(0xFBFF),
|
|
static_cast<int16_t>(0x0000), static_cast<int16_t>(0x7C00),
|
|
static_cast<int16_t>(0xFC00), static_cast<int16_t>(0x0000)}};
|
|
|
|
const struct ExpRes_32I exp_res_fexdo_d[] = {
|
|
{bit_cast<int32_t>(0x7F800000), bit_cast<int32_t>(0x7F7FC99E),
|
|
bit_cast<int32_t>(0x7F800000), bit_cast<int32_t>(0xC49A4000)},
|
|
{bit_cast<int32_t>(0xC21BAE14), bit_cast<int32_t>(0xFF800000),
|
|
bit_cast<int32_t>(0x0082AB1E), bit_cast<int32_t>(0x000BFA5A)},
|
|
{bit_cast<int32_t>(0x7673B164), bit_cast<int32_t>(0xFB13653D),
|
|
bit_cast<int32_t>(0x80000000), bit_cast<int32_t>(0x00000000)},
|
|
{bit_cast<int32_t>(0x000002CA), bit_cast<int32_t>(0x80000000),
|
|
bit_cast<int32_t>(0x80000001), bit_cast<int32_t>(0x00000001)},
|
|
{bit_cast<int32_t>(0xFF800000), bit_cast<int32_t>(0x56B5E621),
|
|
bit_cast<int32_t>(0x00000000), bit_cast<int32_t>(0x7F800000)},
|
|
{bit_cast<int32_t>(0xF673B164), bit_cast<int32_t>(0x7B13653D),
|
|
bit_cast<int32_t>(0x0000042E), bit_cast<int32_t>(0x00000000)}};
|
|
|
|
#define TEST_FEXDO_H(instruction, src, exp_res) \
|
|
run_msa_3rf(reinterpret_cast<const struct TestCaseMsa3RF*>(src), \
|
|
reinterpret_cast<const struct ExpectedResult_MSA3RF*>(exp_res), \
|
|
[](MacroAssembler& assm) { __ instruction(w2, w0, w1); });
|
|
|
|
#define TEST_FEXDO_W(instruction, src, exp_res) \
|
|
run_msa_3rf(reinterpret_cast<const struct TestCaseMsa3RF*>(src), \
|
|
reinterpret_cast<const struct ExpectedResult_MSA3RF*>(exp_res), \
|
|
[](MacroAssembler& assm) { __ instruction(w2, w0, w1); });
|
|
|
|
for (uint64_t i = 0; i < arraysize(tc_w); i++) {
|
|
TEST_FEXDO_H(fexdo_h, &tc_w[i], &exp_res_fexdo_w[i])
|
|
}
|
|
|
|
for (uint64_t i = 0; i < arraysize(tc_d); i++) {
|
|
TEST_FEXDO_W(fexdo_w, &tc_d[i], &exp_res_fexdo_d[i])
|
|
}
|
|
|
|
#undef TEST_FEXDO_H
|
|
#undef TEST_FEXDO_W
|
|
}
|
|
|
|
TEST(MSA_ftq) {
|
|
if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD))
|
|
return;
|
|
|
|
CcTest::InitializeVM();
|
|
|
|
const float nan_float = std::numeric_limits<float>::quiet_NaN();
|
|
const float inf_float = std::numeric_limits<float>::infinity();
|
|
const double nan_double = std::numeric_limits<double>::quiet_NaN();
|
|
const double inf_double = std::numeric_limits<double>::infinity();
|
|
|
|
const struct TestCaseMsa3RF_F tc_w[] = {
|
|
{1.f, -0.999f, 1.5f, -31e-6, 1e-7, -0.598, 0.0023, -0.f, 0, 0, 0, 0},
|
|
{100.f, -102.f, -1.1f, 1.3f, 0.f, -1.f, 0.9999f, -0.000322, 0, 0, 0, 0},
|
|
{nan_float, inf_float, -inf_float, -nan_float, -1e-40, 3e-44, 8.3e36,
|
|
-0.00003, 0, 0, 0, 0}};
|
|
|
|
const struct TestCaseMsa3RF_D tc_d[] = {
|
|
{1., -0.999, 1.5, -31e-6, 0, 0},
|
|
{1e-7, -0.598, 0.0023, -0.f, 0, 0},
|
|
{100.f, -102.f, -1.1f, 1.3f, 0, 0},
|
|
{0.f, -1.f, 0.9999f, -0.000322, 0, 0},
|
|
{nan_double, inf_double, -inf_double, -nan_double, 0, 0},
|
|
{-3e306, 2e-307, 9e307, 2e-307, 0, 0}};
|
|
|
|
const struct ExpRes_16I exp_res_ftq_w[] = {
|
|
{static_cast<int16_t>(0x0000), static_cast<int16_t>(0xB375),
|
|
static_cast<int16_t>(0x004B), static_cast<int16_t>(0x0000),
|
|
static_cast<int16_t>(0x7FFF), static_cast<int16_t>(0x8021),
|
|
static_cast<int16_t>(0x7FFF), static_cast<int16_t>(0xFFFF)},
|
|
{static_cast<int16_t>(0x0000), static_cast<int16_t>(0x8000),
|
|
static_cast<int16_t>(0x7FFD), static_cast<int16_t>(0xFFF5),
|
|
static_cast<int16_t>(0x7FFF), static_cast<int16_t>(0x8000),
|
|
static_cast<int16_t>(0x8000), static_cast<int16_t>(0x7FFF)},
|
|
{static_cast<int16_t>(0x0000), static_cast<int16_t>(0x0000),
|
|
static_cast<int16_t>(0x7FFF), static_cast<int16_t>(0xFFFF),
|
|
static_cast<int16_t>(0x0000), static_cast<int16_t>(0x7FFF),
|
|
static_cast<int16_t>(0x8000), static_cast<int16_t>(0x0000)}};
|
|
|
|
const struct ExpRes_32I exp_res_ftq_d[] = {
|
|
{bit_cast<int32_t>(0x7FFFFFFF), bit_cast<int32_t>(0xFFFEFBF4),
|
|
bit_cast<int32_t>(0x7FFFFFFF), bit_cast<int32_t>(0x8020C49C)},
|
|
{bit_cast<int32_t>(0x004B5DCC), bit_cast<int32_t>(0x00000000),
|
|
bit_cast<int32_t>(0x000000D7), bit_cast<int32_t>(0xB374BC6A)},
|
|
{bit_cast<int32_t>(0x80000000), bit_cast<int32_t>(0x7FFFFFFF),
|
|
bit_cast<int32_t>(0x7FFFFFFF), bit_cast<int32_t>(0x80000000)},
|
|
{bit_cast<int32_t>(0x7FFCB900), bit_cast<int32_t>(0xFFF572DE),
|
|
bit_cast<int32_t>(0x00000000), bit_cast<int32_t>(0x80000000)},
|
|
{bit_cast<int32_t>(0x80000000), bit_cast<int32_t>(0x00000000),
|
|
bit_cast<int32_t>(0x00000000), bit_cast<int32_t>(0x7FFFFFFF)},
|
|
{bit_cast<int32_t>(0x7FFFFFFF), bit_cast<int32_t>(0x00000000),
|
|
bit_cast<int32_t>(0x80000000), bit_cast<int32_t>(0x00000000)}};
|
|
|
|
#define TEST_FTQ_H(instruction, src, exp_res) \
|
|
run_msa_3rf(reinterpret_cast<const struct TestCaseMsa3RF*>(src), \
|
|
reinterpret_cast<const struct ExpectedResult_MSA3RF*>(exp_res), \
|
|
[](MacroAssembler& assm) { __ instruction(w2, w0, w1); });
|
|
|
|
#define TEST_FTQ_W(instruction, src, exp_res) \
|
|
run_msa_3rf(reinterpret_cast<const struct TestCaseMsa3RF*>(src), \
|
|
reinterpret_cast<const struct ExpectedResult_MSA3RF*>(exp_res), \
|
|
[](MacroAssembler& assm) { __ instruction(w2, w0, w1); });
|
|
|
|
for (uint64_t i = 0; i < arraysize(tc_w); i++) {
|
|
TEST_FTQ_H(ftq_h, &tc_w[i], &exp_res_ftq_w[i])
|
|
}
|
|
|
|
for (uint64_t i = 0; i < arraysize(tc_d); i++) {
|
|
TEST_FTQ_W(ftq_w, &tc_d[i], &exp_res_ftq_d[i])
|
|
}
|
|
|
|
#undef TEST_FTQ_H
|
|
#undef TEST_FTQ_W
|
|
}
|
|
|
|
#undef __
|
|
|
|
} // namespace internal
|
|
} // namespace v8
|