// Copyright 2016 the V8 project authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. #include #include "src/base/bits.h" #include "src/base/overflowing-math.h" #include "src/codegen/assembler-inl.h" #include "test/cctest/cctest.h" #include "test/cctest/compiler/value-helper.h" #include "test/cctest/wasm/wasm-run-utils.h" #include "test/common/wasm/wasm-macro-gen.h" namespace v8 { namespace internal { namespace wasm { namespace test_run_wasm_simd { namespace { using DoubleUnOp = double (*)(double); using DoubleBinOp = double (*)(double, double); using DoubleCompareOp = int64_t (*)(double, double); using FloatUnOp = float (*)(float); using FloatBinOp = float (*)(float, float); using FloatCompareOp = int (*)(float, float); using Int64UnOp = int64_t (*)(int64_t); using Int64BinOp = int64_t (*)(int64_t, int64_t); using Int64ShiftOp = int64_t (*)(int64_t, int); using Int32UnOp = int32_t (*)(int32_t); using Int32BinOp = int32_t (*)(int32_t, int32_t); using Int32CompareOp = int (*)(int32_t, int32_t); using Int32ShiftOp = int32_t (*)(int32_t, int); using Int16UnOp = int16_t (*)(int16_t); using Int16BinOp = int16_t (*)(int16_t, int16_t); using Int16CompareOp = int (*)(int16_t, int16_t); using Int16ShiftOp = int16_t (*)(int16_t, int); using Int8UnOp = int8_t (*)(int8_t); using Int8BinOp = int8_t (*)(int8_t, int8_t); using Int8CompareOp = int (*)(int8_t, int8_t); using Int8ShiftOp = int8_t (*)(int8_t, int); #define WASM_SIMD_TEST(name) \ void RunWasm_##name##_Impl(LowerSimd lower_simd, \ ExecutionTier execution_tier); \ TEST(RunWasm_##name##_turbofan) { \ EXPERIMENTAL_FLAG_SCOPE(simd); \ RunWasm_##name##_Impl(kNoLowerSimd, ExecutionTier::kTurbofan); \ } \ TEST(RunWasm_##name##_interpreter) { \ EXPERIMENTAL_FLAG_SCOPE(simd); \ RunWasm_##name##_Impl(kNoLowerSimd, ExecutionTier::kInterpreter); \ } \ TEST(RunWasm_##name##_simd_lowered) { \ EXPERIMENTAL_FLAG_SCOPE(simd); \ RunWasm_##name##_Impl(kLowerSimd, ExecutionTier::kTurbofan); \ } \ void RunWasm_##name##_Impl(LowerSimd lower_simd, ExecutionTier execution_tier) // Generic expected value functions. template ::value>::type> T Negate(T a) { return -a; } // For signed integral types, use base::AddWithWraparound. template ::value>::type> T Add(T a, T b) { return a + b; } // For signed integral types, use base::SubWithWraparound. template ::value>::type> T Sub(T a, T b) { return a - b; } // For signed integral types, use base::MulWithWraparound. template ::value>::type> T Mul(T a, T b) { return a * b; } template T Minimum(T a, T b) { return a <= b ? a : b; } template T Maximum(T a, T b) { return a >= b ? a : b; } template T UnsignedMinimum(T a, T b) { using UnsignedT = typename std::make_unsigned::type; return static_cast(a) <= static_cast(b) ? a : b; } template T UnsignedMaximum(T a, T b) { using UnsignedT = typename std::make_unsigned::type; return static_cast(a) >= static_cast(b) ? a : b; } int Equal(float a, float b) { return a == b ? -1 : 0; } template T Equal(T a, T b) { return a == b ? -1 : 0; } int NotEqual(float a, float b) { return a != b ? -1 : 0; } template T NotEqual(T a, T b) { return a != b ? -1 : 0; } int Less(float a, float b) { return a < b ? -1 : 0; } template T Less(T a, T b) { return a < b ? -1 : 0; } int LessEqual(float a, float b) { return a <= b ? -1 : 0; } template T LessEqual(T a, T b) { return a <= b ? -1 : 0; } int Greater(float a, float b) { return a > b ? -1 : 0; } template T Greater(T a, T b) { return a > b ? -1 : 0; } int GreaterEqual(float a, float b) { return a >= b ? -1 : 0; } template T GreaterEqual(T a, T b) { return a >= b ? -1 : 0; } template T UnsignedLess(T a, T b) { using UnsignedT = typename std::make_unsigned::type; return static_cast(a) < static_cast(b) ? -1 : 0; } template T UnsignedLessEqual(T a, T b) { using UnsignedT = typename std::make_unsigned::type; return static_cast(a) <= static_cast(b) ? -1 : 0; } template T UnsignedGreater(T a, T b) { using UnsignedT = typename std::make_unsigned::type; return static_cast(a) > static_cast(b) ? -1 : 0; } template T UnsignedGreaterEqual(T a, T b) { using UnsignedT = typename std::make_unsigned::type; return static_cast(a) >= static_cast(b) ? -1 : 0; } template T LogicalShiftLeft(T a, int shift) { using UnsignedT = typename std::make_unsigned::type; return static_cast(a) << shift; } template T LogicalShiftRight(T a, int shift) { using UnsignedT = typename std::make_unsigned::type; return static_cast(a) >> shift; } template T Clamp(int64_t value) { static_assert(sizeof(int64_t) > sizeof(T), "T must be int32_t or smaller"); int64_t min = static_cast(std::numeric_limits::min()); int64_t max = static_cast(std::numeric_limits::max()); int64_t clamped = std::max(min, std::min(max, value)); return static_cast(clamped); } template int64_t Widen(T value) { static_assert(sizeof(int64_t) > sizeof(T), "T must be int32_t or smaller"); return static_cast(value); } template int64_t UnsignedWiden(T value) { static_assert(sizeof(int64_t) > sizeof(T), "T must be int32_t or smaller"); using UnsignedT = typename std::make_unsigned::type; return static_cast(static_cast(value)); } template T Narrow(int64_t value) { return Clamp(value); } template T UnsignedNarrow(int64_t value) { static_assert(sizeof(int64_t) > sizeof(T), "T must be int32_t or smaller"); using UnsignedT = typename std::make_unsigned::type; return static_cast(Clamp(value & 0xFFFFFFFFu)); } template T AddSaturate(T a, T b) { return Clamp(Widen(a) + Widen(b)); } template T SubSaturate(T a, T b) { return Clamp(Widen(a) - Widen(b)); } template T UnsignedAddSaturate(T a, T b) { using UnsignedT = typename std::make_unsigned::type; return Clamp(UnsignedWiden(a) + UnsignedWiden(b)); } template T UnsignedSubSaturate(T a, T b) { using UnsignedT = typename std::make_unsigned::type; return Clamp(UnsignedWiden(a) - UnsignedWiden(b)); } template T And(T a, T b) { return a & b; } template T Or(T a, T b) { return a | b; } template T Xor(T a, T b) { return a ^ b; } template T Not(T a) { return ~a; } template T LogicalNot(T a) { return a == 0 ? -1 : 0; } template T Sqrt(T a) { return std::sqrt(a); } #if V8_TARGET_ARCH_X64 // only used for F64x2 tests below int64_t Equal(double a, double b) { return a == b ? -1 : 0; } int64_t NotEqual(double a, double b) { return a != b ? -1 : 0; } int64_t Greater(double a, double b) { return a > b ? -1 : 0; } int64_t GreaterEqual(double a, double b) { return a >= b ? -1 : 0; } int64_t Less(double a, double b) { return a < b ? -1 : 0; } int64_t LessEqual(double a, double b) { return a <= b ? -1 : 0; } #endif // V8_TARGET_ARCH_X64 } // namespace #define WASM_SIMD_CHECK_LANE(TYPE, value, LANE_TYPE, lane_value, lane_index) \ WASM_IF(WASM_##LANE_TYPE##_NE(WASM_GET_LOCAL(lane_value), \ WASM_SIMD_##TYPE##_EXTRACT_LANE( \ lane_index, WASM_GET_LOCAL(value))), \ WASM_RETURN1(WASM_ZERO)) #define TO_BYTE(val) static_cast(val) #define WASM_SIMD_OP(op) kSimdPrefix, TO_BYTE(op) #define WASM_SIMD_SPLAT(Type, ...) __VA_ARGS__, WASM_SIMD_OP(kExpr##Type##Splat) #define WASM_SIMD_UNOP(op, x) x, WASM_SIMD_OP(op) #define WASM_SIMD_BINOP(op, x, y) x, y, WASM_SIMD_OP(op) #define WASM_SIMD_SHIFT_OP(op, shift, x) x, WASM_SIMD_OP(op), TO_BYTE(shift) #define WASM_SIMD_CONCAT_OP(op, bytes, x, y) \ x, y, WASM_SIMD_OP(op), TO_BYTE(bytes) #define WASM_SIMD_SELECT(format, x, y, z) x, y, z, WASM_SIMD_OP(kExprS128Select) #define WASM_SIMD_F64x2_SPLAT(x) WASM_SIMD_SPLAT(F64x2, x) #define WASM_SIMD_F64x2_EXTRACT_LANE(lane, x) \ x, WASM_SIMD_OP(kExprF64x2ExtractLane), TO_BYTE(lane) #define WASM_SIMD_F64x2_REPLACE_LANE(lane, x, y) \ x, y, WASM_SIMD_OP(kExprF64x2ReplaceLane), TO_BYTE(lane) #define WASM_SIMD_F32x4_SPLAT(x) WASM_SIMD_SPLAT(F32x4, x) #define WASM_SIMD_F32x4_EXTRACT_LANE(lane, x) \ x, WASM_SIMD_OP(kExprF32x4ExtractLane), TO_BYTE(lane) #define WASM_SIMD_F32x4_REPLACE_LANE(lane, x, y) \ x, y, WASM_SIMD_OP(kExprF32x4ReplaceLane), TO_BYTE(lane) #define WASM_SIMD_I64x2_SPLAT(x) WASM_SIMD_SPLAT(I64x2, x) #define WASM_SIMD_I64x2_EXTRACT_LANE(lane, x) \ x, WASM_SIMD_OP(kExprI64x2ExtractLane), TO_BYTE(lane) #define WASM_SIMD_I64x2_REPLACE_LANE(lane, x, y) \ x, y, WASM_SIMD_OP(kExprI64x2ReplaceLane), TO_BYTE(lane) #define WASM_SIMD_I32x4_SPLAT(x) WASM_SIMD_SPLAT(I32x4, x) #define WASM_SIMD_I32x4_EXTRACT_LANE(lane, x) \ x, WASM_SIMD_OP(kExprI32x4ExtractLane), TO_BYTE(lane) #define WASM_SIMD_I32x4_REPLACE_LANE(lane, x, y) \ x, y, WASM_SIMD_OP(kExprI32x4ReplaceLane), TO_BYTE(lane) #define WASM_SIMD_I16x8_SPLAT(x) WASM_SIMD_SPLAT(I16x8, x) #define WASM_SIMD_I16x8_EXTRACT_LANE(lane, x) \ x, WASM_SIMD_OP(kExprI16x8ExtractLane), TO_BYTE(lane) #define WASM_SIMD_I16x8_REPLACE_LANE(lane, x, y) \ x, y, WASM_SIMD_OP(kExprI16x8ReplaceLane), TO_BYTE(lane) #define WASM_SIMD_I8x16_SPLAT(x) WASM_SIMD_SPLAT(I8x16, x) #define WASM_SIMD_I8x16_EXTRACT_LANE(lane, x) \ x, WASM_SIMD_OP(kExprI8x16ExtractLane), TO_BYTE(lane) #define WASM_SIMD_I8x16_REPLACE_LANE(lane, x, y) \ x, y, WASM_SIMD_OP(kExprI8x16ReplaceLane), TO_BYTE(lane) #define WASM_SIMD_S8x16_SHUFFLE_OP(opcode, m, x, y) \ x, y, WASM_SIMD_OP(opcode), TO_BYTE(m[0]), TO_BYTE(m[1]), TO_BYTE(m[2]), \ TO_BYTE(m[3]), TO_BYTE(m[4]), TO_BYTE(m[5]), TO_BYTE(m[6]), \ TO_BYTE(m[7]), TO_BYTE(m[8]), TO_BYTE(m[9]), TO_BYTE(m[10]), \ TO_BYTE(m[11]), TO_BYTE(m[12]), TO_BYTE(m[13]), TO_BYTE(m[14]), \ TO_BYTE(m[15]) #define WASM_SIMD_LOAD_MEM(index) \ index, WASM_SIMD_OP(kExprS128LoadMem), ZERO_ALIGNMENT, ZERO_OFFSET #define WASM_SIMD_STORE_MEM(index, val) \ index, val, WASM_SIMD_OP(kExprS128StoreMem), ZERO_ALIGNMENT, ZERO_OFFSET // Runs tests of compiled code, using the interpreter as a reference. #define WASM_SIMD_COMPILED_TEST(name) \ void RunWasm_##name##_Impl(LowerSimd lower_simd, \ ExecutionTier execution_tier); \ TEST(RunWasm_##name##_turbofan) { \ EXPERIMENTAL_FLAG_SCOPE(simd); \ RunWasm_##name##_Impl(kNoLowerSimd, ExecutionTier::kTurbofan); \ } \ TEST(RunWasm_##name##_simd_lowered) { \ EXPERIMENTAL_FLAG_SCOPE(simd); \ RunWasm_##name##_Impl(kLowerSimd, ExecutionTier::kTurbofan); \ } \ void RunWasm_##name##_Impl(LowerSimd lower_simd, ExecutionTier execution_tier) // The macro below disables tests lowering for certain nodes where the simd // lowering doesn't work correctly. Early return here if the CPU does not // support SIMD as the graph will be implicitly lowered in that case. #define WASM_SIMD_TEST_NO_LOWERING(name) \ void RunWasm_##name##_Impl(LowerSimd lower_simd, \ ExecutionTier execution_tier); \ TEST(RunWasm_##name##_turbofan) { \ if (!CpuFeatures::SupportsWasmSimd128()) return; \ EXPERIMENTAL_FLAG_SCOPE(simd); \ RunWasm_##name##_Impl(kNoLowerSimd, ExecutionTier::kTurbofan); \ } \ TEST(RunWasm_##name##_interpreter) { \ EXPERIMENTAL_FLAG_SCOPE(simd); \ RunWasm_##name##_Impl(kNoLowerSimd, ExecutionTier::kInterpreter); \ } \ void RunWasm_##name##_Impl(LowerSimd lower_simd, ExecutionTier execution_tier) // Returns true if the platform can represent the result. template bool PlatformCanRepresent(T x) { #if V8_TARGET_ARCH_ARM return std::fpclassify(x) != FP_SUBNORMAL; #else return true; #endif } // Returns true for very small and very large numbers. We skip these test // values for the approximation instructions, which don't work at the extremes. bool IsExtreme(float x) { float abs_x = std::fabs(x); const float kSmallFloatThreshold = 1.0e-32f; const float kLargeFloatThreshold = 1.0e32f; return abs_x != 0.0f && // 0 or -0 are fine. (abs_x < kSmallFloatThreshold || abs_x > kLargeFloatThreshold); } WASM_SIMD_TEST(F32x4Splat) { WasmRunner r(execution_tier, lower_simd); // Set up a global to hold output vector. float* g = r.builder().AddGlobal(kWasmS128); byte param1 = 0; BUILD(r, WASM_SET_GLOBAL(0, WASM_SIMD_F32x4_SPLAT(WASM_GET_LOCAL(param1))), WASM_ONE); FOR_FLOAT32_INPUTS(x) { r.Call(x); float expected = x; for (int i = 0; i < 4; i++) { float actual = ReadLittleEndianValue(&g[i]); if (std::isnan(expected)) { CHECK(std::isnan(actual)); } else { CHECK_EQ(actual, expected); } } } } WASM_SIMD_TEST(F32x4ReplaceLane) { WasmRunner r(execution_tier, lower_simd); // Set up a global to hold input/output vector. float* g = r.builder().AddGlobal(kWasmS128); // Build function to replace each lane with its (FP) index. byte temp1 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_F32x4_SPLAT(WASM_F32(3.14159f))), WASM_SET_LOCAL(temp1, WASM_SIMD_F32x4_REPLACE_LANE( 0, WASM_GET_LOCAL(temp1), WASM_F32(0.0f))), WASM_SET_LOCAL(temp1, WASM_SIMD_F32x4_REPLACE_LANE( 1, WASM_GET_LOCAL(temp1), WASM_F32(1.0f))), WASM_SET_LOCAL(temp1, WASM_SIMD_F32x4_REPLACE_LANE( 2, WASM_GET_LOCAL(temp1), WASM_F32(2.0f))), WASM_SET_GLOBAL(0, WASM_SIMD_F32x4_REPLACE_LANE( 3, WASM_GET_LOCAL(temp1), WASM_F32(3.0f))), WASM_ONE); r.Call(); for (int i = 0; i < 4; i++) { CHECK_EQ(static_cast(i), ReadLittleEndianValue(&g[i])); } } // Tests both signed and unsigned conversion. // v8:8425 tracks this test being enabled in the interpreter. WASM_SIMD_COMPILED_TEST(F32x4ConvertI32x4) { WasmRunner r(execution_tier, lower_simd); // Create two output vectors to hold signed and unsigned results. float* g0 = r.builder().AddGlobal(kWasmS128); float* g1 = r.builder().AddGlobal(kWasmS128); // Build fn to splat test value, perform conversions, and write the results. byte value = 0; byte temp1 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_I32x4_SPLAT(WASM_GET_LOCAL(value))), WASM_SET_GLOBAL( 0, WASM_SIMD_UNOP(kExprF32x4SConvertI32x4, WASM_GET_LOCAL(temp1))), WASM_SET_GLOBAL( 1, WASM_SIMD_UNOP(kExprF32x4UConvertI32x4, WASM_GET_LOCAL(temp1))), WASM_ONE); FOR_INT32_INPUTS(x) { r.Call(x); float expected_signed = static_cast(x); float expected_unsigned = static_cast(static_cast(x)); for (int i = 0; i < 4; i++) { CHECK_EQ(expected_signed, ReadLittleEndianValue(&g0[i])); CHECK_EQ(expected_unsigned, ReadLittleEndianValue(&g1[i])); } } } bool IsSameNan(float expected, float actual) { // Sign is non-deterministic. uint32_t expected_bits = bit_cast(expected) & ~0x80000000; uint32_t actual_bits = bit_cast(actual) & ~0x80000000; // Some implementations convert signaling NaNs to quiet NaNs. return (expected_bits == actual_bits) || ((expected_bits | 0x00400000) == actual_bits); } bool IsCanonical(float actual) { uint32_t actual_bits = bit_cast(actual); // Canonical NaN has quiet bit and no payload. return (actual_bits & 0xFFC00000) == actual_bits; } void CheckFloatResult(float x, float y, float expected, float actual, bool exact = true) { if (std::isnan(expected)) { CHECK(std::isnan(actual)); if (std::isnan(x) && IsSameNan(x, actual)) return; if (std::isnan(y) && IsSameNan(y, actual)) return; if (IsSameNan(expected, actual)) return; if (IsCanonical(actual)) return; // This is expected to assert; it's useful for debugging. CHECK_EQ(bit_cast(expected), bit_cast(actual)); } else { if (exact) { CHECK_EQ(expected, actual); // The sign of 0's must match. CHECK_EQ(std::signbit(expected), std::signbit(actual)); return; } // Otherwise, perform an approximate equality test. First check for // equality to handle +/-Infinity where approximate equality doesn't work. if (expected == actual) return; // 1% error allows all platforms to pass easily. constexpr float kApproximationError = 0.01f; float abs_error = std::abs(expected) * kApproximationError, min = expected - abs_error, max = expected + abs_error; CHECK_LE(min, actual); CHECK_GE(max, actual); } } // Test some values not included in the float inputs from value_helper. These // tests are useful for opcodes that are synthesized during code gen, like Min // and Max on ia32 and x64. static constexpr uint32_t nan_test_array[] = { // Bit patterns of quiet NaNs and signaling NaNs, with or without // additional payload. 0x7FC00000, 0xFFC00000, 0x7FFFFFFF, 0x7F800000, 0xFF800000, 0x7F876543, 0xFF876543, // Both Infinities. 0x7F800000, 0xFF800000, // Some "normal" numbers, 1 and -1. 0x3F800000, 0xBF800000}; #define FOR_FLOAT32_NAN_INPUTS(i) \ for (size_t i = 0; i < arraysize(nan_test_array); ++i) void RunF32x4UnOpTest(ExecutionTier execution_tier, LowerSimd lower_simd, WasmOpcode opcode, FloatUnOp expected_op, bool exact = true) { WasmRunner r(execution_tier, lower_simd); // Global to hold output. float* g = r.builder().AddGlobal(kWasmS128); // Build fn to splat test value, perform unop, and write the result. byte value = 0; byte temp1 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_F32x4_SPLAT(WASM_GET_LOCAL(value))), WASM_SET_GLOBAL(0, WASM_SIMD_UNOP(opcode, WASM_GET_LOCAL(temp1))), WASM_ONE); FOR_FLOAT32_INPUTS(x) { if (!PlatformCanRepresent(x)) continue; // Extreme values have larger errors so skip them for approximation tests. if (!exact && IsExtreme(x)) continue; float expected = expected_op(x); if (!PlatformCanRepresent(expected)) continue; r.Call(x); for (int i = 0; i < 4; i++) { float actual = ReadLittleEndianValue(&g[i]); CheckFloatResult(x, x, expected, actual, exact); } } FOR_FLOAT32_NAN_INPUTS(i) { float x = bit_cast(nan_test_array[i]); if (!PlatformCanRepresent(x)) continue; // Extreme values have larger errors so skip them for approximation tests. if (!exact && IsExtreme(x)) continue; float expected = expected_op(x); if (!PlatformCanRepresent(expected)) continue; r.Call(x); for (int i = 0; i < 4; i++) { float actual = ReadLittleEndianValue(&g[i]); CheckFloatResult(x, x, expected, actual, exact); } } } WASM_SIMD_TEST(F32x4Abs) { RunF32x4UnOpTest(execution_tier, lower_simd, kExprF32x4Abs, std::abs); } WASM_SIMD_TEST(F32x4Neg) { RunF32x4UnOpTest(execution_tier, lower_simd, kExprF32x4Neg, Negate); } WASM_SIMD_TEST(F32x4RecipApprox) { RunF32x4UnOpTest(execution_tier, lower_simd, kExprF32x4RecipApprox, base::Recip, false /* !exact */); } WASM_SIMD_TEST(F32x4RecipSqrtApprox) { RunF32x4UnOpTest(execution_tier, lower_simd, kExprF32x4RecipSqrtApprox, base::RecipSqrt, false /* !exact */); } void RunF32x4BinOpTest(ExecutionTier execution_tier, LowerSimd lower_simd, WasmOpcode opcode, FloatBinOp expected_op) { WasmRunner r(execution_tier, lower_simd); // Global to hold output. float* g = r.builder().AddGlobal(kWasmS128); // Build fn to splat test values, perform binop, and write the result. byte value1 = 0, value2 = 1; byte temp1 = r.AllocateLocal(kWasmS128); byte temp2 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_F32x4_SPLAT(WASM_GET_LOCAL(value1))), WASM_SET_LOCAL(temp2, WASM_SIMD_F32x4_SPLAT(WASM_GET_LOCAL(value2))), WASM_SET_GLOBAL(0, WASM_SIMD_BINOP(opcode, WASM_GET_LOCAL(temp1), WASM_GET_LOCAL(temp2))), WASM_ONE); FOR_FLOAT32_INPUTS(x) { if (!PlatformCanRepresent(x)) continue; FOR_FLOAT32_INPUTS(y) { if (!PlatformCanRepresent(y)) continue; float expected = expected_op(x, y); if (!PlatformCanRepresent(expected)) continue; r.Call(x, y); for (int i = 0; i < 4; i++) { float actual = ReadLittleEndianValue(&g[i]); CheckFloatResult(x, y, expected, actual, true /* exact */); } } } FOR_FLOAT32_NAN_INPUTS(i) { float x = bit_cast(nan_test_array[i]); if (!PlatformCanRepresent(x)) continue; FOR_FLOAT32_NAN_INPUTS(j) { float y = bit_cast(nan_test_array[j]); if (!PlatformCanRepresent(y)) continue; float expected = expected_op(x, y); if (!PlatformCanRepresent(expected)) continue; r.Call(x, y); for (int i = 0; i < 4; i++) { float actual = ReadLittleEndianValue(&g[i]); CheckFloatResult(x, y, expected, actual, true /* exact */); } } } } #undef FOR_FLOAT32_NAN_INPUTS WASM_SIMD_TEST(F32x4Add) { RunF32x4BinOpTest(execution_tier, lower_simd, kExprF32x4Add, Add); } WASM_SIMD_TEST(F32x4Sub) { RunF32x4BinOpTest(execution_tier, lower_simd, kExprF32x4Sub, Sub); } WASM_SIMD_TEST(F32x4Mul) { RunF32x4BinOpTest(execution_tier, lower_simd, kExprF32x4Mul, Mul); } // v8:8425 tracks this test being enabled in the interpreter. WASM_SIMD_COMPILED_TEST(F32x4Min) { RunF32x4BinOpTest(execution_tier, lower_simd, kExprF32x4Min, JSMin); } // v8:8425 tracks this test being enabled in the interpreter. WASM_SIMD_COMPILED_TEST(F32x4Max) { RunF32x4BinOpTest(execution_tier, lower_simd, kExprF32x4Max, JSMax); } void RunF32x4CompareOpTest(ExecutionTier execution_tier, LowerSimd lower_simd, WasmOpcode opcode, FloatCompareOp expected_op) { WasmRunner r(execution_tier, lower_simd); // Set up global to hold mask output. int32_t* g = r.builder().AddGlobal(kWasmS128); // Build fn to splat test values, perform compare op, and write the result. byte value1 = 0, value2 = 1; byte temp1 = r.AllocateLocal(kWasmS128); byte temp2 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_F32x4_SPLAT(WASM_GET_LOCAL(value1))), WASM_SET_LOCAL(temp2, WASM_SIMD_F32x4_SPLAT(WASM_GET_LOCAL(value2))), WASM_SET_GLOBAL(0, WASM_SIMD_BINOP(opcode, WASM_GET_LOCAL(temp1), WASM_GET_LOCAL(temp2))), WASM_ONE); FOR_FLOAT32_INPUTS(x) { if (!PlatformCanRepresent(x)) continue; FOR_FLOAT32_INPUTS(y) { if (!PlatformCanRepresent(y)) continue; float diff = x - y; // Model comparison as subtraction. if (!PlatformCanRepresent(diff)) continue; r.Call(x, y); int32_t expected = expected_op(x, y); for (int i = 0; i < 4; i++) { CHECK_EQ(expected, ReadLittleEndianValue(&g[i])); } } } } WASM_SIMD_TEST(F32x4Eq) { RunF32x4CompareOpTest(execution_tier, lower_simd, kExprF32x4Eq, Equal); } WASM_SIMD_TEST(F32x4Ne) { RunF32x4CompareOpTest(execution_tier, lower_simd, kExprF32x4Ne, NotEqual); } WASM_SIMD_TEST(F32x4Gt) { RunF32x4CompareOpTest(execution_tier, lower_simd, kExprF32x4Gt, Greater); } WASM_SIMD_TEST(F32x4Ge) { RunF32x4CompareOpTest(execution_tier, lower_simd, kExprF32x4Ge, GreaterEqual); } WASM_SIMD_TEST(F32x4Lt) { RunF32x4CompareOpTest(execution_tier, lower_simd, kExprF32x4Lt, Less); } WASM_SIMD_TEST(F32x4Le) { RunF32x4CompareOpTest(execution_tier, lower_simd, kExprF32x4Le, LessEqual); } #if V8_TARGET_ARCH_X64 WASM_SIMD_TEST_NO_LOWERING(F64x2Splat) { WasmRunner r(execution_tier, lower_simd); // Set up a global to hold output vector. double* g = r.builder().AddGlobal(kWasmS128); byte param1 = 0; BUILD(r, WASM_SET_GLOBAL(0, WASM_SIMD_F64x2_SPLAT(WASM_GET_LOCAL(param1))), WASM_ONE); FOR_FLOAT64_INPUTS(x) { r.Call(x); double expected = x; for (int i = 0; i < 2; i++) { double actual = ReadLittleEndianValue(&g[i]); if (std::isnan(expected)) { CHECK(std::isnan(actual)); } else { CHECK_EQ(actual, expected); } } } } WASM_SIMD_TEST_NO_LOWERING(F64x2ExtractLaneWithI64x2) { WasmRunner r(execution_tier, lower_simd); BUILD(r, WASM_IF_ELSE_L( WASM_F64_EQ(WASM_SIMD_F64x2_EXTRACT_LANE( 0, WASM_SIMD_I64x2_SPLAT(WASM_I64V(1e15))), WASM_F64_REINTERPRET_I64(WASM_I64V(1e15))), WASM_I64V(1), WASM_I64V(0))); CHECK_EQ(1, r.Call()); } WASM_SIMD_TEST_NO_LOWERING(F64x2ExtractLane) { WasmRunner r(execution_tier, lower_simd); byte param1 = 0; byte temp1 = r.AllocateLocal(kWasmF64); byte temp2 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_F64x2_EXTRACT_LANE( 0, WASM_SIMD_F64x2_SPLAT(WASM_GET_LOCAL(param1)))), WASM_SET_LOCAL(temp2, WASM_SIMD_F64x2_SPLAT(WASM_GET_LOCAL(temp1))), WASM_SIMD_F64x2_EXTRACT_LANE(1, WASM_GET_LOCAL(temp2))); FOR_FLOAT64_INPUTS(x) { double actual = r.Call(x); double expected = x; if (std::isnan(expected)) { CHECK(std::isnan(actual)); } else { CHECK_EQ(actual, expected); } } } WASM_SIMD_TEST_NO_LOWERING(F64x2ReplaceLane) { WasmRunner r(execution_tier, lower_simd); // Set up a global to hold input/output vector. double* g = r.builder().AddGlobal(kWasmS128); // Build function to replace each lane with its (FP) index. byte temp1 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_F64x2_SPLAT(WASM_F64(1e100))), WASM_SET_LOCAL(temp1, WASM_SIMD_F64x2_REPLACE_LANE( 0, WASM_GET_LOCAL(temp1), WASM_F64(0.0f))), WASM_SET_GLOBAL(0, WASM_SIMD_F64x2_REPLACE_LANE( 1, WASM_GET_LOCAL(temp1), WASM_F64(1.0f))), WASM_ONE); r.Call(); for (int i = 0; i < 2; i++) { CHECK_EQ(static_cast(i), ReadLittleEndianValue(&g[i])); } } void RunF64x2CompareOpTest(ExecutionTier execution_tier, LowerSimd lower_simd, WasmOpcode opcode, DoubleCompareOp expected_op) { WasmRunner r(execution_tier, lower_simd); // Set up global to hold mask output. int64_t* g = r.builder().AddGlobal(kWasmS128); // Build fn to splat test values, perform compare op, and write the result. byte value1 = 0, value2 = 1; byte temp1 = r.AllocateLocal(kWasmS128); byte temp2 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_F64x2_SPLAT(WASM_GET_LOCAL(value1))), WASM_SET_LOCAL(temp2, WASM_SIMD_F64x2_SPLAT(WASM_GET_LOCAL(value2))), WASM_SET_GLOBAL(0, WASM_SIMD_BINOP(opcode, WASM_GET_LOCAL(temp1), WASM_GET_LOCAL(temp2))), WASM_ONE); FOR_FLOAT64_INPUTS(x) { if (!PlatformCanRepresent(x)) continue; FOR_FLOAT64_INPUTS(y) { if (!PlatformCanRepresent(y)) continue; double diff = x - y; // Model comparison as subtraction. if (!PlatformCanRepresent(diff)) continue; r.Call(x, y); int64_t expected = expected_op(x, y); for (int i = 0; i < 2; i++) { CHECK_EQ(expected, ReadLittleEndianValue(&g[i])); } } } } WASM_SIMD_TEST_NO_LOWERING(F64x2Eq) { RunF64x2CompareOpTest(execution_tier, lower_simd, kExprF64x2Eq, Equal); } WASM_SIMD_TEST_NO_LOWERING(F64x2Ne) { RunF64x2CompareOpTest(execution_tier, lower_simd, kExprF64x2Ne, NotEqual); } WASM_SIMD_TEST_NO_LOWERING(F64x2Gt) { RunF64x2CompareOpTest(execution_tier, lower_simd, kExprF64x2Gt, Greater); } WASM_SIMD_TEST_NO_LOWERING(F64x2Ge) { RunF64x2CompareOpTest(execution_tier, lower_simd, kExprF64x2Ge, GreaterEqual); } WASM_SIMD_TEST_NO_LOWERING(F64x2Lt) { RunF64x2CompareOpTest(execution_tier, lower_simd, kExprF64x2Lt, Less); } WASM_SIMD_TEST_NO_LOWERING(F64x2Le) { RunF64x2CompareOpTest(execution_tier, lower_simd, kExprF64x2Le, LessEqual); } bool IsSameNan(double expected, double actual) { // Sign is non-deterministic. uint64_t expected_bits = bit_cast(expected) & ~0x8000000000000000; uint64_t actual_bits = bit_cast(actual) & ~0x8000000000000000; // Some implementations convert signaling NaNs to quiet NaNs. return (expected_bits == actual_bits) || ((expected_bits | 0x0008000000000000) == actual_bits); } bool IsCanonical(double actual) { uint64_t actual_bits = bit_cast(actual); // Canonical NaN has quiet bit and no payload. return (actual_bits & 0xFF80000000000000) == actual_bits; } void CheckDoubleResult(double x, double y, double expected, double actual, bool exact = true) { if (std::isnan(expected)) { CHECK(std::isnan(actual)); if (std::isnan(x) && IsSameNan(x, actual)) return; if (std::isnan(y) && IsSameNan(y, actual)) return; if (IsSameNan(expected, actual)) return; if (IsCanonical(actual)) return; // This is expected to assert; it's useful for debugging. CHECK_EQ(bit_cast(expected), bit_cast(actual)); } else { if (exact) { CHECK_EQ(expected, actual); // The sign of 0's must match. CHECK_EQ(std::signbit(expected), std::signbit(actual)); return; } // Otherwise, perform an approximate equality test. First check for // equality to handle +/-Infinity where approximate equality doesn't work. if (expected == actual) return; // 1% error allows all platforms to pass easily. constexpr double kApproximationError = 0.01f; double abs_error = std::abs(expected) * kApproximationError, min = expected - abs_error, max = expected + abs_error; CHECK_LE(min, actual); CHECK_GE(max, actual); } } // Test some values not included in the double inputs from value_helper. These // tests are useful for opcodes that are synthesized during code gen, like Min // and Max on ia32 and x64. static constexpr uint64_t double_nan_test_array[] = { // quiet NaNs, + and - 0x7FF8000000000001, 0xFFF8000000000001, // with payload 0x7FF8000000000011, 0xFFF8000000000011, // signaling NaNs, + and - 0x7FF0000000000001, 0xFFF0000000000001, // with payload 0x7FF0000000000011, 0xFFF0000000000011, // Both Infinities. 0x7FF0000000000000, 0xFFF0000000000000, // Some "normal" numbers, 1 and -1. 0x3FF0000000000000, 0xBFF0000000000000}; #define FOR_FLOAT64_NAN_INPUTS(i) \ for (size_t i = 0; i < arraysize(double_nan_test_array); ++i) void RunF64x2UnOpTest(ExecutionTier execution_tier, LowerSimd lower_simd, WasmOpcode opcode, DoubleUnOp expected_op, bool exact = true) { WasmRunner r(execution_tier, lower_simd); // Global to hold output. double* g = r.builder().AddGlobal(kWasmS128); // Build fn to splat test value, perform unop, and write the result. byte value = 0; byte temp1 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_F64x2_SPLAT(WASM_GET_LOCAL(value))), WASM_SET_GLOBAL(0, WASM_SIMD_UNOP(opcode, WASM_GET_LOCAL(temp1))), WASM_ONE); FOR_FLOAT64_INPUTS(x) { if (!PlatformCanRepresent(x)) continue; // Extreme values have larger errors so skip them for approximation tests. if (!exact && IsExtreme(x)) continue; double expected = expected_op(x); if (!PlatformCanRepresent(expected)) continue; r.Call(x); for (int i = 0; i < 2; i++) { double actual = ReadLittleEndianValue(&g[i]); CheckDoubleResult(x, x, expected, actual, exact); } } FOR_FLOAT64_NAN_INPUTS(i) { double x = bit_cast(double_nan_test_array[i]); if (!PlatformCanRepresent(x)) continue; // Extreme values have larger errors so skip them for approximation tests. if (!exact && IsExtreme(x)) continue; double expected = expected_op(x); if (!PlatformCanRepresent(expected)) continue; r.Call(x); for (int i = 0; i < 2; i++) { double actual = ReadLittleEndianValue(&g[i]); CheckDoubleResult(x, x, expected, actual, exact); } } } WASM_SIMD_TEST_NO_LOWERING(F64x2Abs) { RunF64x2UnOpTest(execution_tier, lower_simd, kExprF64x2Abs, std::abs); } WASM_SIMD_TEST_NO_LOWERING(F64x2Neg) { RunF64x2UnOpTest(execution_tier, lower_simd, kExprF64x2Neg, Negate); } void RunF64x2BinOpTest(ExecutionTier execution_tier, LowerSimd lower_simd, WasmOpcode opcode, DoubleBinOp expected_op) { WasmRunner r(execution_tier, lower_simd); // Global to hold output. double* g = r.builder().AddGlobal(kWasmS128); // Build fn to splat test value, perform binop, and write the result. byte value1 = 0, value2 = 1; byte temp1 = r.AllocateLocal(kWasmS128); byte temp2 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_F64x2_SPLAT(WASM_GET_LOCAL(value1))), WASM_SET_LOCAL(temp2, WASM_SIMD_F64x2_SPLAT(WASM_GET_LOCAL(value2))), WASM_SET_GLOBAL(0, WASM_SIMD_BINOP(opcode, WASM_GET_LOCAL(temp1), WASM_GET_LOCAL(temp2))), WASM_ONE); FOR_FLOAT64_INPUTS(x) { if (!PlatformCanRepresent(x)) continue; FOR_FLOAT64_INPUTS(y) { if (!PlatformCanRepresent(x)) continue; double expected = expected_op(x, y); if (!PlatformCanRepresent(expected)) continue; r.Call(x, y); for (int i = 0; i < 2; i++) { double actual = ReadLittleEndianValue(&g[i]); CheckDoubleResult(x, y, expected, actual, true /* exact */); } } } FOR_FLOAT64_NAN_INPUTS(i) { double x = bit_cast(double_nan_test_array[i]); if (!PlatformCanRepresent(x)) continue; FOR_FLOAT64_NAN_INPUTS(j) { double y = bit_cast(double_nan_test_array[j]); double expected = expected_op(x, y); if (!PlatformCanRepresent(expected)) continue; r.Call(x, y); for (int i = 0; i < 2; i++) { double actual = ReadLittleEndianValue(&g[i]); CheckDoubleResult(x, y, expected, actual, true /* exact */); } } } } WASM_SIMD_TEST_NO_LOWERING(F64x2Add) { RunF64x2BinOpTest(execution_tier, lower_simd, kExprF64x2Add, Add); } WASM_SIMD_TEST_NO_LOWERING(F64x2Sub) { RunF64x2BinOpTest(execution_tier, lower_simd, kExprF64x2Sub, Sub); } WASM_SIMD_TEST_NO_LOWERING(F64x2Mul) { RunF64x2BinOpTest(execution_tier, lower_simd, kExprF64x2Mul, Mul); } #undef FOR_FLOAT64_NAN_INPUTS WASM_SIMD_TEST_NO_LOWERING(I64x2Splat) { WasmRunner r(execution_tier, lower_simd); // Set up a global to hold output vector. int64_t* g = r.builder().AddGlobal(kWasmS128); byte param1 = 0; BUILD(r, WASM_SET_GLOBAL(0, WASM_SIMD_I64x2_SPLAT(WASM_GET_LOCAL(param1))), WASM_ONE); FOR_INT64_INPUTS(x) { r.Call(x); int64_t expected = x; for (int i = 0; i < 2; i++) { int64_t actual = ReadLittleEndianValue(&g[i]); CHECK_EQ(actual, expected); } } } WASM_SIMD_TEST_NO_LOWERING(I64x2ExtractWithF64x2) { WasmRunner r(execution_tier, lower_simd); BUILD(r, WASM_IF_ELSE_L( WASM_I64_EQ(WASM_SIMD_I64x2_EXTRACT_LANE( 0, WASM_SIMD_F64x2_SPLAT(WASM_F64(1e15))), WASM_I64_REINTERPRET_F64(WASM_F64(1e15))), WASM_I64V(1), WASM_I64V(0))); CHECK_EQ(1, r.Call()); } WASM_SIMD_TEST_NO_LOWERING(I64x2ReplaceLane) { WasmRunner r(execution_tier, lower_simd); // Set up a global to hold input/output vector. int64_t* g = r.builder().AddGlobal(kWasmS128); // Build function to replace each lane with its index. byte temp1 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_I64x2_SPLAT(WASM_I64V(-1))), WASM_SET_LOCAL(temp1, WASM_SIMD_I64x2_REPLACE_LANE( 0, WASM_GET_LOCAL(temp1), WASM_I64V(0))), WASM_SET_GLOBAL(0, WASM_SIMD_I64x2_REPLACE_LANE( 1, WASM_GET_LOCAL(temp1), WASM_I64V(1))), WASM_ONE); r.Call(); for (int64_t i = 0; i < 2; i++) { CHECK_EQ(i, ReadLittleEndianValue(&g[i])); } } void RunI64x2UnOpTest(ExecutionTier execution_tier, LowerSimd lower_simd, WasmOpcode opcode, Int64UnOp expected_op) { WasmRunner r(execution_tier, lower_simd); // Global to hold output. int64_t* g = r.builder().AddGlobal(kWasmS128); // Build fn to splat test value, perform unop, and write the result. byte value = 0; byte temp1 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_I64x2_SPLAT(WASM_GET_LOCAL(value))), WASM_SET_GLOBAL(0, WASM_SIMD_UNOP(opcode, WASM_GET_LOCAL(temp1))), WASM_ONE); FOR_INT64_INPUTS(x) { r.Call(x); int64_t expected = expected_op(x); for (int i = 0; i < 2; i++) { CHECK_EQ(expected, ReadLittleEndianValue(&g[i])); } } } WASM_SIMD_TEST_NO_LOWERING(I64x2Neg) { RunI64x2UnOpTest(execution_tier, lower_simd, kExprI64x2Neg, base::NegateWithWraparound); } void RunI64x2BinOpTest(ExecutionTier execution_tier, LowerSimd lower_simd, WasmOpcode opcode, Int64BinOp expected_op) { WasmRunner r(execution_tier, lower_simd); // Global to hold output. int64_t* g = r.builder().AddGlobal(kWasmS128); // Build fn to splat test values, perform binop, and write the result. byte value1 = 0, value2 = 1; byte temp1 = r.AllocateLocal(kWasmS128); byte temp2 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_I64x2_SPLAT(WASM_GET_LOCAL(value1))), WASM_SET_LOCAL(temp2, WASM_SIMD_I64x2_SPLAT(WASM_GET_LOCAL(value2))), WASM_SET_GLOBAL(0, WASM_SIMD_BINOP(opcode, WASM_GET_LOCAL(temp1), WASM_GET_LOCAL(temp2))), WASM_ONE); FOR_INT64_INPUTS(x) { FOR_INT64_INPUTS(y) { r.Call(x, y); int64_t expected = expected_op(x, y); for (int i = 0; i < 2; i++) { CHECK_EQ(expected, ReadLittleEndianValue(&g[i])); } } } } WASM_SIMD_TEST_NO_LOWERING(I64x2Add) { RunI64x2BinOpTest(execution_tier, lower_simd, kExprI64x2Add, base::AddWithWraparound); } WASM_SIMD_TEST_NO_LOWERING(I64x2Sub) { RunI64x2BinOpTest(execution_tier, lower_simd, kExprI64x2Sub, base::SubWithWraparound); } WASM_SIMD_TEST_NO_LOWERING(I64x2Mul) { RunI64x2BinOpTest(execution_tier, lower_simd, kExprI64x2Mul, base::MulWithWraparound); } WASM_SIMD_TEST_NO_LOWERING(I64x2MinS) { RunI64x2BinOpTest(execution_tier, lower_simd, kExprI64x2MinS, Minimum); } WASM_SIMD_TEST_NO_LOWERING(I64x2MaxS) { RunI64x2BinOpTest(execution_tier, lower_simd, kExprI64x2MaxS, Maximum); } WASM_SIMD_TEST_NO_LOWERING(I64x2MinU) { RunI64x2BinOpTest(execution_tier, lower_simd, kExprI64x2MinU, UnsignedMinimum); } WASM_SIMD_TEST_NO_LOWERING(I64x2MaxU) { RunI64x2BinOpTest(execution_tier, lower_simd, kExprI64x2MaxU, UnsignedMaximum); } WASM_SIMD_TEST_NO_LOWERING(I64x2Eq) { RunI64x2BinOpTest(execution_tier, lower_simd, kExprI64x2Eq, Equal); } WASM_SIMD_TEST_NO_LOWERING(I64x2Ne) { RunI64x2BinOpTest(execution_tier, lower_simd, kExprI64x2Ne, NotEqual); } WASM_SIMD_TEST_NO_LOWERING(I64x2LtS) { RunI64x2BinOpTest(execution_tier, lower_simd, kExprI64x2LtS, Less); } WASM_SIMD_TEST_NO_LOWERING(I64x2LeS) { RunI64x2BinOpTest(execution_tier, lower_simd, kExprI64x2LeS, LessEqual); } WASM_SIMD_TEST_NO_LOWERING(I64x2GtS) { RunI64x2BinOpTest(execution_tier, lower_simd, kExprI64x2GtS, Greater); } WASM_SIMD_TEST_NO_LOWERING(I64x2GeS) { RunI64x2BinOpTest(execution_tier, lower_simd, kExprI64x2GeS, GreaterEqual); } WASM_SIMD_TEST_NO_LOWERING(I64x2LtU) { RunI64x2BinOpTest(execution_tier, lower_simd, kExprI64x2LtU, UnsignedLess); } WASM_SIMD_TEST_NO_LOWERING(I64x2LeU) { RunI64x2BinOpTest(execution_tier, lower_simd, kExprI64x2LeU, UnsignedLessEqual); } WASM_SIMD_TEST_NO_LOWERING(I64x2GtU) { RunI64x2BinOpTest(execution_tier, lower_simd, kExprI64x2GtU, UnsignedGreater); } WASM_SIMD_TEST_NO_LOWERING(I64x2GeU) { RunI64x2BinOpTest(execution_tier, lower_simd, kExprI64x2GeU, UnsignedGreaterEqual); } void RunI64x2ShiftOpTest(ExecutionTier execution_tier, LowerSimd lower_simd, WasmOpcode opcode, Int64ShiftOp expected_op) { for (int shift = 1; shift < 64; shift++) { WasmRunner r(execution_tier, lower_simd); int64_t* g = r.builder().AddGlobal(kWasmS128); byte value = 0; byte simd1 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(simd1, WASM_SIMD_I64x2_SPLAT(WASM_GET_LOCAL(value))), WASM_SET_GLOBAL( 0, WASM_SIMD_SHIFT_OP(opcode, shift, WASM_GET_LOCAL(simd1))), WASM_ONE); FOR_INT64_INPUTS(x) { r.Call(x); int64_t expected = expected_op(x, shift); for (int i = 0; i < 2; i++) { CHECK_EQ(expected, ReadLittleEndianValue(&g[i])); } } } } WASM_SIMD_TEST_NO_LOWERING(I64x2Shl) { RunI64x2ShiftOpTest(execution_tier, lower_simd, kExprI64x2Shl, LogicalShiftLeft); } WASM_SIMD_TEST_NO_LOWERING(I64x2ShrS) { RunI64x2ShiftOpTest(execution_tier, lower_simd, kExprI64x2ShrS, ArithmeticShiftRight); } WASM_SIMD_TEST_NO_LOWERING(I64x2ShrU) { RunI64x2ShiftOpTest(execution_tier, lower_simd, kExprI64x2ShrU, LogicalShiftRight); } #endif // V8_TARGET_ARCH_X64 WASM_SIMD_TEST(I32x4Splat) { WasmRunner r(execution_tier, lower_simd); // Set up a global to hold output vector. int32_t* g = r.builder().AddGlobal(kWasmS128); byte param1 = 0; BUILD(r, WASM_SET_GLOBAL(0, WASM_SIMD_I32x4_SPLAT(WASM_GET_LOCAL(param1))), WASM_ONE); FOR_INT32_INPUTS(x) { r.Call(x); int32_t expected = x; for (int i = 0; i < 4; i++) { int32_t actual = ReadLittleEndianValue(&g[i]); CHECK_EQ(actual, expected); } } } WASM_SIMD_TEST(I32x4ReplaceLane) { WasmRunner r(execution_tier, lower_simd); // Set up a global to hold input/output vector. int32_t* g = r.builder().AddGlobal(kWasmS128); // Build function to replace each lane with its index. byte temp1 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_I32x4_SPLAT(WASM_I32V(-1))), WASM_SET_LOCAL(temp1, WASM_SIMD_I32x4_REPLACE_LANE( 0, WASM_GET_LOCAL(temp1), WASM_I32V(0))), WASM_SET_LOCAL(temp1, WASM_SIMD_I32x4_REPLACE_LANE( 1, WASM_GET_LOCAL(temp1), WASM_I32V(1))), WASM_SET_LOCAL(temp1, WASM_SIMD_I32x4_REPLACE_LANE( 2, WASM_GET_LOCAL(temp1), WASM_I32V(2))), WASM_SET_GLOBAL(0, WASM_SIMD_I32x4_REPLACE_LANE( 3, WASM_GET_LOCAL(temp1), WASM_I32V(3))), WASM_ONE); r.Call(); for (int32_t i = 0; i < 4; i++) { CHECK_EQ(i, ReadLittleEndianValue(&g[i])); } } WASM_SIMD_TEST(I16x8Splat) { WasmRunner r(execution_tier, lower_simd); // Set up a global to hold output vector. int16_t* g = r.builder().AddGlobal(kWasmS128); byte param1 = 0; BUILD(r, WASM_SET_GLOBAL(0, WASM_SIMD_I16x8_SPLAT(WASM_GET_LOCAL(param1))), WASM_ONE); FOR_INT16_INPUTS(x) { r.Call(x); int16_t expected = x; for (int i = 0; i < 8; i++) { int16_t actual = ReadLittleEndianValue(&g[i]); CHECK_EQ(actual, expected); } } } WASM_SIMD_TEST(I16x8ReplaceLane) { WasmRunner r(execution_tier, lower_simd); // Set up a global to hold input/output vector. int16_t* g = r.builder().AddGlobal(kWasmS128); // Build function to replace each lane with its index. byte temp1 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_I16x8_SPLAT(WASM_I32V(-1))), WASM_SET_LOCAL(temp1, WASM_SIMD_I16x8_REPLACE_LANE( 0, WASM_GET_LOCAL(temp1), WASM_I32V(0))), WASM_SET_LOCAL(temp1, WASM_SIMD_I16x8_REPLACE_LANE( 1, WASM_GET_LOCAL(temp1), WASM_I32V(1))), WASM_SET_LOCAL(temp1, WASM_SIMD_I16x8_REPLACE_LANE( 2, WASM_GET_LOCAL(temp1), WASM_I32V(2))), WASM_SET_LOCAL(temp1, WASM_SIMD_I16x8_REPLACE_LANE( 3, WASM_GET_LOCAL(temp1), WASM_I32V(3))), WASM_SET_LOCAL(temp1, WASM_SIMD_I16x8_REPLACE_LANE( 4, WASM_GET_LOCAL(temp1), WASM_I32V(4))), WASM_SET_LOCAL(temp1, WASM_SIMD_I16x8_REPLACE_LANE( 5, WASM_GET_LOCAL(temp1), WASM_I32V(5))), WASM_SET_LOCAL(temp1, WASM_SIMD_I16x8_REPLACE_LANE( 6, WASM_GET_LOCAL(temp1), WASM_I32V(6))), WASM_SET_GLOBAL(0, WASM_SIMD_I16x8_REPLACE_LANE( 7, WASM_GET_LOCAL(temp1), WASM_I32V(7))), WASM_ONE); r.Call(); for (int16_t i = 0; i < 8; i++) { CHECK_EQ(i, ReadLittleEndianValue(&g[i])); } } WASM_SIMD_TEST(I8x16Splat) { WasmRunner r(execution_tier, lower_simd); // Set up a global to hold output vector. int8_t* g = r.builder().AddGlobal(kWasmS128); byte param1 = 0; BUILD(r, WASM_SET_GLOBAL(0, WASM_SIMD_I8x16_SPLAT(WASM_GET_LOCAL(param1))), WASM_ONE); FOR_INT8_INPUTS(x) { r.Call(x); int8_t expected = x; for (int i = 0; i < 16; i++) { int8_t actual = ReadLittleEndianValue(&g[i]); CHECK_EQ(actual, expected); } } } WASM_SIMD_TEST(I8x16ReplaceLane) { WasmRunner r(execution_tier, lower_simd); // Set up a global to hold input/output vector. int8_t* g = r.builder().AddGlobal(kWasmS128); // Build function to replace each lane with its index. byte temp1 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_I8x16_SPLAT(WASM_I32V(-1))), WASM_SET_LOCAL(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 0, WASM_GET_LOCAL(temp1), WASM_I32V(0))), WASM_SET_LOCAL(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 1, WASM_GET_LOCAL(temp1), WASM_I32V(1))), WASM_SET_LOCAL(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 2, WASM_GET_LOCAL(temp1), WASM_I32V(2))), WASM_SET_LOCAL(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 3, WASM_GET_LOCAL(temp1), WASM_I32V(3))), WASM_SET_LOCAL(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 4, WASM_GET_LOCAL(temp1), WASM_I32V(4))), WASM_SET_LOCAL(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 5, WASM_GET_LOCAL(temp1), WASM_I32V(5))), WASM_SET_LOCAL(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 6, WASM_GET_LOCAL(temp1), WASM_I32V(6))), WASM_SET_LOCAL(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 7, WASM_GET_LOCAL(temp1), WASM_I32V(7))), WASM_SET_LOCAL(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 8, WASM_GET_LOCAL(temp1), WASM_I32V(8))), WASM_SET_LOCAL(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 9, WASM_GET_LOCAL(temp1), WASM_I32V(9))), WASM_SET_LOCAL(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 10, WASM_GET_LOCAL(temp1), WASM_I32V(10))), WASM_SET_LOCAL(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 11, WASM_GET_LOCAL(temp1), WASM_I32V(11))), WASM_SET_LOCAL(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 12, WASM_GET_LOCAL(temp1), WASM_I32V(12))), WASM_SET_LOCAL(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 13, WASM_GET_LOCAL(temp1), WASM_I32V(13))), WASM_SET_LOCAL(temp1, WASM_SIMD_I8x16_REPLACE_LANE( 14, WASM_GET_LOCAL(temp1), WASM_I32V(14))), WASM_SET_GLOBAL(0, WASM_SIMD_I8x16_REPLACE_LANE( 15, WASM_GET_LOCAL(temp1), WASM_I32V(15))), WASM_ONE); r.Call(); for (int8_t i = 0; i < 16; i++) { CHECK_EQ(i, ReadLittleEndianValue(&g[i])); } } // Use doubles to ensure exact conversion. int32_t ConvertToInt(double val, bool unsigned_integer) { if (std::isnan(val)) return 0; if (unsigned_integer) { if (val < 0) return 0; if (val > kMaxUInt32) return kMaxUInt32; return static_cast(val); } else { if (val < kMinInt) return kMinInt; if (val > kMaxInt) return kMaxInt; return static_cast(val); } } // Tests both signed and unsigned conversion. WASM_SIMD_TEST(I32x4ConvertF32x4) { WasmRunner r(execution_tier, lower_simd); // Create two output vectors to hold signed and unsigned results. int32_t* g0 = r.builder().AddGlobal(kWasmS128); int32_t* g1 = r.builder().AddGlobal(kWasmS128); // Build fn to splat test value, perform conversions, and write the results. byte value = 0; byte temp1 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_F32x4_SPLAT(WASM_GET_LOCAL(value))), WASM_SET_GLOBAL( 0, WASM_SIMD_UNOP(kExprI32x4SConvertF32x4, WASM_GET_LOCAL(temp1))), WASM_SET_GLOBAL( 1, WASM_SIMD_UNOP(kExprI32x4UConvertF32x4, WASM_GET_LOCAL(temp1))), WASM_ONE); FOR_FLOAT32_INPUTS(x) { if (!PlatformCanRepresent(x)) continue; r.Call(x); int32_t expected_signed = ConvertToInt(x, false); int32_t expected_unsigned = ConvertToInt(x, true); for (int i = 0; i < 4; i++) { CHECK_EQ(expected_signed, ReadLittleEndianValue(&g0[i])); CHECK_EQ(expected_unsigned, ReadLittleEndianValue(&g1[i])); } } } // Tests both signed and unsigned conversion from I16x8 (unpacking). WASM_SIMD_TEST(I32x4ConvertI16x8) { WasmRunner r(execution_tier, lower_simd); // Create four output vectors to hold signed and unsigned results. int32_t* g0 = r.builder().AddGlobal(kWasmS128); int32_t* g1 = r.builder().AddGlobal(kWasmS128); int32_t* g2 = r.builder().AddGlobal(kWasmS128); int32_t* g3 = r.builder().AddGlobal(kWasmS128); // Build fn to splat test value, perform conversions, and write the results. byte value = 0; byte temp1 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_I16x8_SPLAT(WASM_GET_LOCAL(value))), WASM_SET_GLOBAL(0, WASM_SIMD_UNOP(kExprI32x4SConvertI16x8High, WASM_GET_LOCAL(temp1))), WASM_SET_GLOBAL(1, WASM_SIMD_UNOP(kExprI32x4SConvertI16x8Low, WASM_GET_LOCAL(temp1))), WASM_SET_GLOBAL(2, WASM_SIMD_UNOP(kExprI32x4UConvertI16x8High, WASM_GET_LOCAL(temp1))), WASM_SET_GLOBAL(3, WASM_SIMD_UNOP(kExprI32x4UConvertI16x8Low, WASM_GET_LOCAL(temp1))), WASM_ONE); FOR_INT16_INPUTS(x) { r.Call(x); int32_t expected_signed = static_cast(Widen(x)); int32_t expected_unsigned = static_cast(UnsignedWiden(x)); for (int i = 0; i < 4; i++) { CHECK_EQ(expected_signed, ReadLittleEndianValue(&g0[i])); CHECK_EQ(expected_signed, ReadLittleEndianValue(&g1[i])); CHECK_EQ(expected_unsigned, ReadLittleEndianValue(&g2[i])); CHECK_EQ(expected_unsigned, ReadLittleEndianValue(&g3[i])); } } } void RunI32x4UnOpTest(ExecutionTier execution_tier, LowerSimd lower_simd, WasmOpcode opcode, Int32UnOp expected_op) { WasmRunner r(execution_tier, lower_simd); // Global to hold output. int32_t* g = r.builder().AddGlobal(kWasmS128); // Build fn to splat test value, perform unop, and write the result. byte value = 0; byte temp1 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_I32x4_SPLAT(WASM_GET_LOCAL(value))), WASM_SET_GLOBAL(0, WASM_SIMD_UNOP(opcode, WASM_GET_LOCAL(temp1))), WASM_ONE); FOR_INT32_INPUTS(x) { r.Call(x); int32_t expected = expected_op(x); for (int i = 0; i < 4; i++) { CHECK_EQ(expected, ReadLittleEndianValue(&g[i])); } } } WASM_SIMD_TEST(I32x4Neg) { RunI32x4UnOpTest(execution_tier, lower_simd, kExprI32x4Neg, base::NegateWithWraparound); } WASM_SIMD_TEST(S128Not) { RunI32x4UnOpTest(execution_tier, lower_simd, kExprS128Not, Not); } void RunI32x4BinOpTest(ExecutionTier execution_tier, LowerSimd lower_simd, WasmOpcode opcode, Int32BinOp expected_op) { WasmRunner r(execution_tier, lower_simd); // Global to hold output. int32_t* g = r.builder().AddGlobal(kWasmS128); // Build fn to splat test values, perform binop, and write the result. byte value1 = 0, value2 = 1; byte temp1 = r.AllocateLocal(kWasmS128); byte temp2 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_I32x4_SPLAT(WASM_GET_LOCAL(value1))), WASM_SET_LOCAL(temp2, WASM_SIMD_I32x4_SPLAT(WASM_GET_LOCAL(value2))), WASM_SET_GLOBAL(0, WASM_SIMD_BINOP(opcode, WASM_GET_LOCAL(temp1), WASM_GET_LOCAL(temp2))), WASM_ONE); FOR_INT32_INPUTS(x) { FOR_INT32_INPUTS(y) { r.Call(x, y); int32_t expected = expected_op(x, y); for (int i = 0; i < 4; i++) { CHECK_EQ(expected, ReadLittleEndianValue(&g[i])); } } } } WASM_SIMD_TEST(I32x4Add) { RunI32x4BinOpTest(execution_tier, lower_simd, kExprI32x4Add, base::AddWithWraparound); } WASM_SIMD_TEST(I32x4Sub) { RunI32x4BinOpTest(execution_tier, lower_simd, kExprI32x4Sub, base::SubWithWraparound); } WASM_SIMD_TEST(I32x4Mul) { RunI32x4BinOpTest(execution_tier, lower_simd, kExprI32x4Mul, base::MulWithWraparound); } WASM_SIMD_TEST(I32x4MinS) { RunI32x4BinOpTest(execution_tier, lower_simd, kExprI32x4MinS, Minimum); } WASM_SIMD_TEST(I32x4MaxS) { RunI32x4BinOpTest(execution_tier, lower_simd, kExprI32x4MaxS, Maximum); } WASM_SIMD_TEST(I32x4MinU) { RunI32x4BinOpTest(execution_tier, lower_simd, kExprI32x4MinU, UnsignedMinimum); } WASM_SIMD_TEST(I32x4MaxU) { RunI32x4BinOpTest(execution_tier, lower_simd, kExprI32x4MaxU, UnsignedMaximum); } WASM_SIMD_TEST(S128And) { RunI32x4BinOpTest(execution_tier, lower_simd, kExprS128And, And); } WASM_SIMD_TEST(S128Or) { RunI32x4BinOpTest(execution_tier, lower_simd, kExprS128Or, Or); } WASM_SIMD_TEST(S128Xor) { RunI32x4BinOpTest(execution_tier, lower_simd, kExprS128Xor, Xor); } WASM_SIMD_TEST(I32x4Eq) { RunI32x4BinOpTest(execution_tier, lower_simd, kExprI32x4Eq, Equal); } WASM_SIMD_TEST(I32x4Ne) { RunI32x4BinOpTest(execution_tier, lower_simd, kExprI32x4Ne, NotEqual); } WASM_SIMD_TEST(I32x4LtS) { RunI32x4BinOpTest(execution_tier, lower_simd, kExprI32x4LtS, Less); } WASM_SIMD_TEST(I32x4LeS) { RunI32x4BinOpTest(execution_tier, lower_simd, kExprI32x4LeS, LessEqual); } WASM_SIMD_TEST(I32x4GtS) { RunI32x4BinOpTest(execution_tier, lower_simd, kExprI32x4GtS, Greater); } WASM_SIMD_TEST(I32x4GeS) { RunI32x4BinOpTest(execution_tier, lower_simd, kExprI32x4GeS, GreaterEqual); } WASM_SIMD_TEST(I32x4LtU) { RunI32x4BinOpTest(execution_tier, lower_simd, kExprI32x4LtU, UnsignedLess); } WASM_SIMD_TEST(I32x4LeU) { RunI32x4BinOpTest(execution_tier, lower_simd, kExprI32x4LeU, UnsignedLessEqual); } WASM_SIMD_TEST(I32x4GtU) { RunI32x4BinOpTest(execution_tier, lower_simd, kExprI32x4GtU, UnsignedGreater); } WASM_SIMD_TEST(I32x4GeU) { RunI32x4BinOpTest(execution_tier, lower_simd, kExprI32x4GeU, UnsignedGreaterEqual); } void RunI32x4ShiftOpTest(ExecutionTier execution_tier, LowerSimd lower_simd, WasmOpcode opcode, Int32ShiftOp expected_op) { for (int shift = 1; shift < 32; shift++) { WasmRunner r(execution_tier, lower_simd); int32_t* g = r.builder().AddGlobal(kWasmS128); byte value = 0; byte simd1 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(simd1, WASM_SIMD_I32x4_SPLAT(WASM_GET_LOCAL(value))), WASM_SET_GLOBAL( 0, WASM_SIMD_SHIFT_OP(opcode, shift, WASM_GET_LOCAL(simd1))), WASM_ONE); FOR_INT32_INPUTS(x) { r.Call(x); float expected = expected_op(x, shift); for (int i = 0; i < 4; i++) { CHECK_EQ(expected, ReadLittleEndianValue(&g[i])); } } } } WASM_SIMD_TEST(I32x4Shl) { RunI32x4ShiftOpTest(execution_tier, lower_simd, kExprI32x4Shl, LogicalShiftLeft); } WASM_SIMD_TEST(I32x4ShrS) { RunI32x4ShiftOpTest(execution_tier, lower_simd, kExprI32x4ShrS, ArithmeticShiftRight); } WASM_SIMD_TEST(I32x4ShrU) { RunI32x4ShiftOpTest(execution_tier, lower_simd, kExprI32x4ShrU, LogicalShiftRight); } // Tests both signed and unsigned conversion from I8x16 (unpacking). WASM_SIMD_TEST(I16x8ConvertI8x16) { WasmRunner r(execution_tier, lower_simd); // Create four output vectors to hold signed and unsigned results. int16_t* g0 = r.builder().AddGlobal(kWasmS128); int16_t* g1 = r.builder().AddGlobal(kWasmS128); int16_t* g2 = r.builder().AddGlobal(kWasmS128); int16_t* g3 = r.builder().AddGlobal(kWasmS128); // Build fn to splat test value, perform conversions, and write the results. byte value = 0; byte temp1 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_I8x16_SPLAT(WASM_GET_LOCAL(value))), WASM_SET_GLOBAL(0, WASM_SIMD_UNOP(kExprI16x8SConvertI8x16High, WASM_GET_LOCAL(temp1))), WASM_SET_GLOBAL(1, WASM_SIMD_UNOP(kExprI16x8SConvertI8x16Low, WASM_GET_LOCAL(temp1))), WASM_SET_GLOBAL(2, WASM_SIMD_UNOP(kExprI16x8UConvertI8x16High, WASM_GET_LOCAL(temp1))), WASM_SET_GLOBAL(3, WASM_SIMD_UNOP(kExprI16x8UConvertI8x16Low, WASM_GET_LOCAL(temp1))), WASM_ONE); FOR_INT8_INPUTS(x) { r.Call(x); int16_t expected_signed = static_cast(Widen(x)); int16_t expected_unsigned = static_cast(UnsignedWiden(x)); for (int i = 0; i < 8; i++) { CHECK_EQ(expected_signed, ReadLittleEndianValue(&g0[i])); CHECK_EQ(expected_signed, ReadLittleEndianValue(&g1[i])); CHECK_EQ(expected_unsigned, ReadLittleEndianValue(&g2[i])); CHECK_EQ(expected_unsigned, ReadLittleEndianValue(&g3[i])); } } } // Tests both signed and unsigned conversion from I32x4 (packing). WASM_SIMD_TEST(I16x8ConvertI32x4) { WasmRunner r(execution_tier, lower_simd); // Create output vectors to hold signed and unsigned results. int16_t* g0 = r.builder().AddGlobal(kWasmS128); int16_t* g1 = r.builder().AddGlobal(kWasmS128); // Build fn to splat test value, perform conversions, and write the results. byte value = 0; byte temp1 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_I32x4_SPLAT(WASM_GET_LOCAL(value))), WASM_SET_GLOBAL( 0, WASM_SIMD_BINOP(kExprI16x8SConvertI32x4, WASM_GET_LOCAL(temp1), WASM_GET_LOCAL(temp1))), WASM_SET_GLOBAL( 1, WASM_SIMD_BINOP(kExprI16x8UConvertI32x4, WASM_GET_LOCAL(temp1), WASM_GET_LOCAL(temp1))), WASM_ONE); FOR_INT32_INPUTS(x) { r.Call(x); int16_t expected_signed = Narrow(x); int16_t expected_unsigned = UnsignedNarrow(x); for (int i = 0; i < 8; i++) { CHECK_EQ(expected_signed, ReadLittleEndianValue(&g0[i])); CHECK_EQ(expected_unsigned, ReadLittleEndianValue(&g1[i])); } } } void RunI16x8UnOpTest(ExecutionTier execution_tier, LowerSimd lower_simd, WasmOpcode opcode, Int16UnOp expected_op) { WasmRunner r(execution_tier, lower_simd); // Global to hold output. int16_t* g = r.builder().AddGlobal(kWasmS128); // Build fn to splat test value, perform unop, and write the result. byte value = 0; byte temp1 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_I16x8_SPLAT(WASM_GET_LOCAL(value))), WASM_SET_GLOBAL(0, WASM_SIMD_UNOP(opcode, WASM_GET_LOCAL(temp1))), WASM_ONE); FOR_INT16_INPUTS(x) { r.Call(x); int16_t expected = expected_op(x); for (int i = 0; i < 8; i++) { CHECK_EQ(expected, ReadLittleEndianValue(&g[i])); } } } WASM_SIMD_TEST(I16x8Neg) { RunI16x8UnOpTest(execution_tier, lower_simd, kExprI16x8Neg, base::NegateWithWraparound); } void RunI16x8BinOpTest(ExecutionTier execution_tier, LowerSimd lower_simd, WasmOpcode opcode, Int16BinOp expected_op) { WasmRunner r(execution_tier, lower_simd); // Global to hold output. int16_t* g = r.builder().AddGlobal(kWasmS128); // Build fn to splat test values, perform binop, and write the result. byte value1 = 0, value2 = 1; byte temp1 = r.AllocateLocal(kWasmS128); byte temp2 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_I16x8_SPLAT(WASM_GET_LOCAL(value1))), WASM_SET_LOCAL(temp2, WASM_SIMD_I16x8_SPLAT(WASM_GET_LOCAL(value2))), WASM_SET_GLOBAL(0, WASM_SIMD_BINOP(opcode, WASM_GET_LOCAL(temp1), WASM_GET_LOCAL(temp2))), WASM_ONE); FOR_INT16_INPUTS(x) { FOR_INT16_INPUTS(y) { r.Call(x, y); int16_t expected = expected_op(x, y); for (int i = 0; i < 8; i++) { CHECK_EQ(expected, ReadLittleEndianValue(&g[i])); } } } } WASM_SIMD_TEST(I16x8Add) { RunI16x8BinOpTest(execution_tier, lower_simd, kExprI16x8Add, base::AddWithWraparound); } WASM_SIMD_TEST(I16x8AddSaturateS) { RunI16x8BinOpTest(execution_tier, lower_simd, kExprI16x8AddSaturateS, AddSaturate); } WASM_SIMD_TEST(I16x8Sub) { RunI16x8BinOpTest(execution_tier, lower_simd, kExprI16x8Sub, base::SubWithWraparound); } WASM_SIMD_TEST(I16x8SubSaturateS) { RunI16x8BinOpTest(execution_tier, lower_simd, kExprI16x8SubSaturateS, SubSaturate); } WASM_SIMD_TEST(I16x8Mul) { RunI16x8BinOpTest(execution_tier, lower_simd, kExprI16x8Mul, base::MulWithWraparound); } WASM_SIMD_TEST(I16x8MinS) { RunI16x8BinOpTest(execution_tier, lower_simd, kExprI16x8MinS, Minimum); } WASM_SIMD_TEST(I16x8MaxS) { RunI16x8BinOpTest(execution_tier, lower_simd, kExprI16x8MaxS, Maximum); } WASM_SIMD_TEST(I16x8AddSaturateU) { RunI16x8BinOpTest(execution_tier, lower_simd, kExprI16x8AddSaturateU, UnsignedAddSaturate); } WASM_SIMD_TEST(I16x8SubSaturateU) { RunI16x8BinOpTest(execution_tier, lower_simd, kExprI16x8SubSaturateU, UnsignedSubSaturate); } WASM_SIMD_TEST(I16x8MinU) { RunI16x8BinOpTest(execution_tier, lower_simd, kExprI16x8MinU, UnsignedMinimum); } WASM_SIMD_TEST(I16x8MaxU) { RunI16x8BinOpTest(execution_tier, lower_simd, kExprI16x8MaxU, UnsignedMaximum); } WASM_SIMD_TEST(I16x8Eq) { RunI16x8BinOpTest(execution_tier, lower_simd, kExprI16x8Eq, Equal); } WASM_SIMD_TEST(I16x8Ne) { RunI16x8BinOpTest(execution_tier, lower_simd, kExprI16x8Ne, NotEqual); } WASM_SIMD_TEST(I16x8LtS) { RunI16x8BinOpTest(execution_tier, lower_simd, kExprI16x8LtS, Less); } WASM_SIMD_TEST(I16x8LeS) { RunI16x8BinOpTest(execution_tier, lower_simd, kExprI16x8LeS, LessEqual); } WASM_SIMD_TEST(I16x8GtS) { RunI16x8BinOpTest(execution_tier, lower_simd, kExprI16x8GtS, Greater); } WASM_SIMD_TEST(I16x8GeS) { RunI16x8BinOpTest(execution_tier, lower_simd, kExprI16x8GeS, GreaterEqual); } WASM_SIMD_TEST(I16x8GtU) { RunI16x8BinOpTest(execution_tier, lower_simd, kExprI16x8GtU, UnsignedGreater); } WASM_SIMD_TEST(I16x8GeU) { RunI16x8BinOpTest(execution_tier, lower_simd, kExprI16x8GeU, UnsignedGreaterEqual); } WASM_SIMD_TEST(I16x8LtU) { RunI16x8BinOpTest(execution_tier, lower_simd, kExprI16x8LtU, UnsignedLess); } WASM_SIMD_TEST(I16x8LeU) { RunI16x8BinOpTest(execution_tier, lower_simd, kExprI16x8LeU, UnsignedLessEqual); } void RunI16x8ShiftOpTest(ExecutionTier execution_tier, LowerSimd lower_simd, WasmOpcode opcode, Int16ShiftOp expected_op) { for (int shift = 1; shift < 16; shift++) { WasmRunner r(execution_tier, lower_simd); int16_t* g = r.builder().AddGlobal(kWasmS128); byte value = 0; byte simd1 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(simd1, WASM_SIMD_I16x8_SPLAT(WASM_GET_LOCAL(value))), WASM_SET_GLOBAL( 0, WASM_SIMD_SHIFT_OP(opcode, shift, WASM_GET_LOCAL(simd1))), WASM_ONE); FOR_INT16_INPUTS(x) { r.Call(x); float expected = expected_op(x, shift); for (int i = 0; i < 8; i++) { CHECK_EQ(expected, ReadLittleEndianValue(&g[i])); } } } } WASM_SIMD_TEST(I16x8Shl) { RunI16x8ShiftOpTest(execution_tier, lower_simd, kExprI16x8Shl, LogicalShiftLeft); } WASM_SIMD_TEST(I16x8ShrS) { RunI16x8ShiftOpTest(execution_tier, lower_simd, kExprI16x8ShrS, ArithmeticShiftRight); } WASM_SIMD_TEST(I16x8ShrU) { RunI16x8ShiftOpTest(execution_tier, lower_simd, kExprI16x8ShrU, LogicalShiftRight); } void RunI8x16UnOpTest(ExecutionTier execution_tier, LowerSimd lower_simd, WasmOpcode opcode, Int8UnOp expected_op) { WasmRunner r(execution_tier, lower_simd); // Global to hold output. int8_t* g = r.builder().AddGlobal(kWasmS128); // Build fn to splat test value, perform unop, and write the result. byte value = 0; byte temp1 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_I8x16_SPLAT(WASM_GET_LOCAL(value))), WASM_SET_GLOBAL(0, WASM_SIMD_UNOP(opcode, WASM_GET_LOCAL(temp1))), WASM_ONE); FOR_INT8_INPUTS(x) { r.Call(x); int8_t expected = expected_op(x); for (int i = 0; i < 16; i++) { CHECK_EQ(expected, ReadLittleEndianValue(&g[i])); } } } WASM_SIMD_TEST(I8x16Neg) { RunI8x16UnOpTest(execution_tier, lower_simd, kExprI8x16Neg, base::NegateWithWraparound); } // Tests both signed and unsigned conversion from I16x8 (packing). WASM_SIMD_TEST(I8x16ConvertI16x8) { WasmRunner r(execution_tier, lower_simd); // Create output vectors to hold signed and unsigned results. int8_t* g0 = r.builder().AddGlobal(kWasmS128); int8_t* g1 = r.builder().AddGlobal(kWasmS128); // Build fn to splat test value, perform conversions, and write the results. byte value = 0; byte temp1 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_I16x8_SPLAT(WASM_GET_LOCAL(value))), WASM_SET_GLOBAL( 0, WASM_SIMD_BINOP(kExprI8x16SConvertI16x8, WASM_GET_LOCAL(temp1), WASM_GET_LOCAL(temp1))), WASM_SET_GLOBAL( 1, WASM_SIMD_BINOP(kExprI8x16UConvertI16x8, WASM_GET_LOCAL(temp1), WASM_GET_LOCAL(temp1))), WASM_ONE); FOR_INT16_INPUTS(x) { r.Call(x); int8_t expected_signed = Narrow(x); int8_t expected_unsigned = UnsignedNarrow(x); for (int i = 0; i < 16; i++) { CHECK_EQ(expected_signed, ReadLittleEndianValue(&g0[i])); CHECK_EQ(expected_unsigned, ReadLittleEndianValue(&g1[i])); } } } void RunI8x16BinOpTest(ExecutionTier execution_tier, LowerSimd lower_simd, WasmOpcode opcode, Int8BinOp expected_op) { WasmRunner r(execution_tier, lower_simd); // Global to hold output. int8_t* g = r.builder().AddGlobal(kWasmS128); // Build fn to splat test values, perform binop, and write the result. byte value1 = 0, value2 = 1; byte temp1 = r.AllocateLocal(kWasmS128); byte temp2 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_I8x16_SPLAT(WASM_GET_LOCAL(value1))), WASM_SET_LOCAL(temp2, WASM_SIMD_I8x16_SPLAT(WASM_GET_LOCAL(value2))), WASM_SET_GLOBAL(0, WASM_SIMD_BINOP(opcode, WASM_GET_LOCAL(temp1), WASM_GET_LOCAL(temp2))), WASM_ONE); FOR_INT8_INPUTS(x) { FOR_INT8_INPUTS(y) { r.Call(x, y); int8_t expected = expected_op(x, y); for (int i = 0; i < 16; i++) { CHECK_EQ(expected, ReadLittleEndianValue(&g[i])); } } } } WASM_SIMD_TEST(I8x16Add) { RunI8x16BinOpTest(execution_tier, lower_simd, kExprI8x16Add, base::AddWithWraparound); } WASM_SIMD_TEST(I8x16AddSaturateS) { RunI8x16BinOpTest(execution_tier, lower_simd, kExprI8x16AddSaturateS, AddSaturate); } WASM_SIMD_TEST(I8x16Sub) { RunI8x16BinOpTest(execution_tier, lower_simd, kExprI8x16Sub, base::SubWithWraparound); } WASM_SIMD_TEST(I8x16SubSaturateS) { RunI8x16BinOpTest(execution_tier, lower_simd, kExprI8x16SubSaturateS, SubSaturate); } WASM_SIMD_TEST(I8x16MinS) { RunI8x16BinOpTest(execution_tier, lower_simd, kExprI8x16MinS, Minimum); } WASM_SIMD_TEST(I8x16MaxS) { RunI8x16BinOpTest(execution_tier, lower_simd, kExprI8x16MaxS, Maximum); } WASM_SIMD_TEST(I8x16AddSaturateU) { RunI8x16BinOpTest(execution_tier, lower_simd, kExprI8x16AddSaturateU, UnsignedAddSaturate); } WASM_SIMD_TEST(I8x16SubSaturateU) { RunI8x16BinOpTest(execution_tier, lower_simd, kExprI8x16SubSaturateU, UnsignedSubSaturate); } WASM_SIMD_TEST(I8x16MinU) { RunI8x16BinOpTest(execution_tier, lower_simd, kExprI8x16MinU, UnsignedMinimum); } WASM_SIMD_TEST(I8x16MaxU) { RunI8x16BinOpTest(execution_tier, lower_simd, kExprI8x16MaxU, UnsignedMaximum); } WASM_SIMD_TEST(I8x16Eq) { RunI8x16BinOpTest(execution_tier, lower_simd, kExprI8x16Eq, Equal); } WASM_SIMD_TEST(I8x16Ne) { RunI8x16BinOpTest(execution_tier, lower_simd, kExprI8x16Ne, NotEqual); } WASM_SIMD_TEST(I8x16GtS) { RunI8x16BinOpTest(execution_tier, lower_simd, kExprI8x16GtS, Greater); } WASM_SIMD_TEST(I8x16GeS) { RunI8x16BinOpTest(execution_tier, lower_simd, kExprI8x16GeS, GreaterEqual); } WASM_SIMD_TEST(I8x16LtS) { RunI8x16BinOpTest(execution_tier, lower_simd, kExprI8x16LtS, Less); } WASM_SIMD_TEST(I8x16LeS) { RunI8x16BinOpTest(execution_tier, lower_simd, kExprI8x16LeS, LessEqual); } WASM_SIMD_TEST(I8x16GtU) { RunI8x16BinOpTest(execution_tier, lower_simd, kExprI8x16GtU, UnsignedGreater); } WASM_SIMD_TEST(I8x16GeU) { RunI8x16BinOpTest(execution_tier, lower_simd, kExprI8x16GeU, UnsignedGreaterEqual); } WASM_SIMD_TEST(I8x16LtU) { RunI8x16BinOpTest(execution_tier, lower_simd, kExprI8x16LtU, UnsignedLess); } WASM_SIMD_TEST(I8x16LeU) { RunI8x16BinOpTest(execution_tier, lower_simd, kExprI8x16LeU, UnsignedLessEqual); } WASM_SIMD_TEST(I8x16Mul) { RunI8x16BinOpTest(execution_tier, lower_simd, kExprI8x16Mul, base::MulWithWraparound); } void RunI8x16ShiftOpTest(ExecutionTier execution_tier, LowerSimd lower_simd, WasmOpcode opcode, Int8ShiftOp expected_op) { for (int shift = 1; shift < 8; shift++) { WasmRunner r(execution_tier, lower_simd); int8_t* g = r.builder().AddGlobal(kWasmS128); byte value = 0; byte simd1 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(simd1, WASM_SIMD_I8x16_SPLAT(WASM_GET_LOCAL(value))), WASM_SET_GLOBAL( 0, WASM_SIMD_SHIFT_OP(opcode, shift, WASM_GET_LOCAL(simd1))), WASM_ONE); FOR_INT8_INPUTS(x) { r.Call(x); float expected = expected_op(x, shift); for (int i = 0; i < 16; i++) { CHECK_EQ(expected, ReadLittleEndianValue(&g[i])); } } } } WASM_SIMD_TEST(I8x16Shl) { RunI8x16ShiftOpTest(execution_tier, lower_simd, kExprI8x16Shl, LogicalShiftLeft); } WASM_SIMD_TEST(I8x16ShrS) { RunI8x16ShiftOpTest(execution_tier, lower_simd, kExprI8x16ShrS, ArithmeticShiftRight); } WASM_SIMD_TEST(I8x16ShrU) { RunI8x16ShiftOpTest(execution_tier, lower_simd, kExprI8x16ShrU, LogicalShiftRight); } // Test Select by making a mask where the 0th and 3rd lanes are true and the // rest false, and comparing for non-equality with zero to convert to a boolean // vector. #define WASM_SIMD_SELECT_TEST(format) \ WASM_SIMD_TEST(S##format##Select) { \ WasmRunner r(execution_tier, lower_simd); \ byte val1 = 0; \ byte val2 = 1; \ byte src1 = r.AllocateLocal(kWasmS128); \ byte src2 = r.AllocateLocal(kWasmS128); \ byte zero = r.AllocateLocal(kWasmS128); \ byte mask = r.AllocateLocal(kWasmS128); \ BUILD(r, \ WASM_SET_LOCAL(src1, \ WASM_SIMD_I##format##_SPLAT(WASM_GET_LOCAL(val1))), \ WASM_SET_LOCAL(src2, \ WASM_SIMD_I##format##_SPLAT(WASM_GET_LOCAL(val2))), \ WASM_SET_LOCAL(zero, WASM_SIMD_I##format##_SPLAT(WASM_ZERO)), \ WASM_SET_LOCAL(mask, WASM_SIMD_I##format##_REPLACE_LANE( \ 1, WASM_GET_LOCAL(zero), WASM_I32V(-1))), \ WASM_SET_LOCAL(mask, WASM_SIMD_I##format##_REPLACE_LANE( \ 2, WASM_GET_LOCAL(mask), WASM_I32V(-1))), \ WASM_SET_LOCAL( \ mask, \ WASM_SIMD_SELECT( \ format, WASM_GET_LOCAL(src1), WASM_GET_LOCAL(src2), \ WASM_SIMD_BINOP(kExprI##format##Ne, WASM_GET_LOCAL(mask), \ WASM_GET_LOCAL(zero)))), \ WASM_SIMD_CHECK_LANE(I##format, mask, I32, val2, 0), \ WASM_SIMD_CHECK_LANE(I##format, mask, I32, val1, 1), \ WASM_SIMD_CHECK_LANE(I##format, mask, I32, val1, 2), \ WASM_SIMD_CHECK_LANE(I##format, mask, I32, val2, 3), WASM_ONE); \ \ CHECK_EQ(1, r.Call(0x12, 0x34)); \ } WASM_SIMD_SELECT_TEST(32x4) WASM_SIMD_SELECT_TEST(16x8) WASM_SIMD_SELECT_TEST(8x16) // Test Select by making a mask where the 0th and 3rd lanes are non-zero and the // rest 0. The mask is not the result of a comparison op. #define WASM_SIMD_NON_CANONICAL_SELECT_TEST(format) \ WASM_SIMD_TEST_NO_LOWERING(S##format##NonCanonicalSelect) { \ WasmRunner r(execution_tier, \ lower_simd); \ byte val1 = 0; \ byte val2 = 1; \ byte combined = 2; \ byte src1 = r.AllocateLocal(kWasmS128); \ byte src2 = r.AllocateLocal(kWasmS128); \ byte zero = r.AllocateLocal(kWasmS128); \ byte mask = r.AllocateLocal(kWasmS128); \ BUILD(r, \ WASM_SET_LOCAL(src1, \ WASM_SIMD_I##format##_SPLAT(WASM_GET_LOCAL(val1))), \ WASM_SET_LOCAL(src2, \ WASM_SIMD_I##format##_SPLAT(WASM_GET_LOCAL(val2))), \ WASM_SET_LOCAL(zero, WASM_SIMD_I##format##_SPLAT(WASM_ZERO)), \ WASM_SET_LOCAL(mask, WASM_SIMD_I##format##_REPLACE_LANE( \ 1, WASM_GET_LOCAL(zero), WASM_I32V(0xF))), \ WASM_SET_LOCAL(mask, WASM_SIMD_I##format##_REPLACE_LANE( \ 2, WASM_GET_LOCAL(mask), WASM_I32V(0xF))), \ WASM_SET_LOCAL(mask, WASM_SIMD_SELECT(format, WASM_GET_LOCAL(src1), \ WASM_GET_LOCAL(src2), \ WASM_GET_LOCAL(mask))), \ WASM_SIMD_CHECK_LANE(I##format, mask, I32, val2, 0), \ WASM_SIMD_CHECK_LANE(I##format, mask, I32, combined, 1), \ WASM_SIMD_CHECK_LANE(I##format, mask, I32, combined, 2), \ WASM_SIMD_CHECK_LANE(I##format, mask, I32, val2, 3), WASM_ONE); \ \ CHECK_EQ(1, r.Call(0x12, 0x34, 0x32)); \ } WASM_SIMD_NON_CANONICAL_SELECT_TEST(32x4) WASM_SIMD_NON_CANONICAL_SELECT_TEST(16x8) WASM_SIMD_NON_CANONICAL_SELECT_TEST(8x16) // Test binary ops with two lane test patterns, all lanes distinct. template void RunBinaryLaneOpTest( ExecutionTier execution_tier, LowerSimd lower_simd, WasmOpcode simd_op, const std::array& expected) { WasmRunner r(execution_tier, lower_simd); // Set up two test patterns as globals, e.g. [0, 1, 2, 3] and [4, 5, 6, 7]. T* src0 = r.builder().AddGlobal(kWasmS128); T* src1 = r.builder().AddGlobal(kWasmS128); static const int kElems = kSimd128Size / sizeof(T); for (int i = 0; i < kElems; i++) { WriteLittleEndianValue(&src0[i], i); WriteLittleEndianValue(&src1[i], kElems + i); } if (simd_op == kExprS8x16Shuffle) { BUILD(r, WASM_SET_GLOBAL(0, WASM_SIMD_S8x16_SHUFFLE_OP(simd_op, expected, WASM_GET_GLOBAL(0), WASM_GET_GLOBAL(1))), WASM_ONE); } else { BUILD(r, WASM_SET_GLOBAL(0, WASM_SIMD_BINOP(simd_op, WASM_GET_GLOBAL(0), WASM_GET_GLOBAL(1))), WASM_ONE); } CHECK_EQ(1, r.Call()); for (size_t i = 0; i < expected.size(); i++) { CHECK_EQ(ReadLittleEndianValue(&src0[i]), expected[i]); } } WASM_SIMD_TEST(I32x4AddHoriz) { // Inputs are [0 1 2 3] and [4 5 6 7]. RunBinaryLaneOpTest(execution_tier, lower_simd, kExprI32x4AddHoriz, {{1, 5, 9, 13}}); } WASM_SIMD_TEST(I16x8AddHoriz) { // Inputs are [0 1 2 3 4 5 6 7] and [8 9 10 11 12 13 14 15]. RunBinaryLaneOpTest(execution_tier, lower_simd, kExprI16x8AddHoriz, {{1, 5, 9, 13, 17, 21, 25, 29}}); } WASM_SIMD_TEST(F32x4AddHoriz) { // Inputs are [0.0f 1.0f 2.0f 3.0f] and [4.0f 5.0f 6.0f 7.0f]. RunBinaryLaneOpTest(execution_tier, lower_simd, kExprF32x4AddHoriz, {{1.0f, 5.0f, 9.0f, 13.0f}}); } // Test shuffle ops. void RunShuffleOpTest(ExecutionTier execution_tier, LowerSimd lower_simd, WasmOpcode simd_op, const std::array& shuffle) { // Test the original shuffle. RunBinaryLaneOpTest(execution_tier, lower_simd, simd_op, shuffle); // Test a non-canonical (inputs reversed) version of the shuffle. std::array other_shuffle(shuffle); for (size_t i = 0; i < shuffle.size(); ++i) other_shuffle[i] ^= kSimd128Size; RunBinaryLaneOpTest(execution_tier, lower_simd, simd_op, other_shuffle); // Test the swizzle (one-operand) version of the shuffle. std::array swizzle(shuffle); for (size_t i = 0; i < shuffle.size(); ++i) swizzle[i] &= (kSimd128Size - 1); RunBinaryLaneOpTest(execution_tier, lower_simd, simd_op, swizzle); // Test the non-canonical swizzle (one-operand) version of the shuffle. std::array other_swizzle(shuffle); for (size_t i = 0; i < shuffle.size(); ++i) other_swizzle[i] |= kSimd128Size; RunBinaryLaneOpTest(execution_tier, lower_simd, simd_op, other_swizzle); } #define SHUFFLE_LIST(V) \ V(S128Identity) \ V(S32x4Dup) \ V(S32x4ZipLeft) \ V(S32x4ZipRight) \ V(S32x4UnzipLeft) \ V(S32x4UnzipRight) \ V(S32x4TransposeLeft) \ V(S32x4TransposeRight) \ V(S32x2Reverse) \ V(S32x4Irregular) \ V(S16x8Dup) \ V(S16x8ZipLeft) \ V(S16x8ZipRight) \ V(S16x8UnzipLeft) \ V(S16x8UnzipRight) \ V(S16x8TransposeLeft) \ V(S16x8TransposeRight) \ V(S16x4Reverse) \ V(S16x2Reverse) \ V(S16x8Irregular) \ V(S8x16Dup) \ V(S8x16ZipLeft) \ V(S8x16ZipRight) \ V(S8x16UnzipLeft) \ V(S8x16UnzipRight) \ V(S8x16TransposeLeft) \ V(S8x16TransposeRight) \ V(S8x8Reverse) \ V(S8x4Reverse) \ V(S8x2Reverse) \ V(S8x16Irregular) enum ShuffleKey { #define SHUFFLE_ENUM_VALUE(Name) k##Name, SHUFFLE_LIST(SHUFFLE_ENUM_VALUE) #undef SHUFFLE_ENUM_VALUE kNumShuffleKeys }; using Shuffle = std::array; using ShuffleMap = std::map; ShuffleMap test_shuffles = { {kS128Identity, {{16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31}}}, {kS32x4Dup, {{16, 17, 18, 19, 16, 17, 18, 19, 16, 17, 18, 19, 16, 17, 18, 19}}}, {kS32x4ZipLeft, {{0, 1, 2, 3, 16, 17, 18, 19, 4, 5, 6, 7, 20, 21, 22, 23}}}, {kS32x4ZipRight, {{8, 9, 10, 11, 24, 25, 26, 27, 12, 13, 14, 15, 28, 29, 30, 31}}}, {kS32x4UnzipLeft, {{0, 1, 2, 3, 8, 9, 10, 11, 16, 17, 18, 19, 24, 25, 26, 27}}}, {kS32x4UnzipRight, {{4, 5, 6, 7, 12, 13, 14, 15, 20, 21, 22, 23, 28, 29, 30, 31}}}, {kS32x4TransposeLeft, {{0, 1, 2, 3, 16, 17, 18, 19, 8, 9, 10, 11, 24, 25, 26, 27}}}, {kS32x4TransposeRight, {{4, 5, 6, 7, 20, 21, 22, 23, 12, 13, 14, 15, 28, 29, 30, 31}}}, {kS32x2Reverse, // swizzle only {{4, 5, 6, 7, 0, 1, 2, 3, 12, 13, 14, 15, 8, 9, 10, 11}}}, {kS32x4Irregular, {{0, 1, 2, 3, 16, 17, 18, 19, 16, 17, 18, 19, 20, 21, 22, 23}}}, {kS16x8Dup, {{18, 19, 18, 19, 18, 19, 18, 19, 18, 19, 18, 19, 18, 19, 18, 19}}}, {kS16x8ZipLeft, {{0, 1, 16, 17, 2, 3, 18, 19, 4, 5, 20, 21, 6, 7, 22, 23}}}, {kS16x8ZipRight, {{8, 9, 24, 25, 10, 11, 26, 27, 12, 13, 28, 29, 14, 15, 30, 31}}}, {kS16x8UnzipLeft, {{0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}}}, {kS16x8UnzipRight, {{2, 3, 6, 7, 10, 11, 14, 15, 18, 19, 22, 23, 26, 27, 30, 31}}}, {kS16x8TransposeLeft, {{0, 1, 16, 17, 4, 5, 20, 21, 8, 9, 24, 25, 12, 13, 28, 29}}}, {kS16x8TransposeRight, {{2, 3, 18, 19, 6, 7, 22, 23, 10, 11, 26, 27, 14, 15, 30, 31}}}, {kS16x4Reverse, // swizzle only {{6, 7, 4, 5, 2, 3, 0, 1, 14, 15, 12, 13, 10, 11, 8, 9}}}, {kS16x2Reverse, // swizzle only {{2, 3, 0, 1, 6, 7, 4, 5, 10, 11, 8, 9, 14, 15, 12, 13}}}, {kS16x8Irregular, {{0, 1, 16, 17, 16, 17, 0, 1, 4, 5, 20, 21, 6, 7, 22, 23}}}, {kS8x16Dup, {{19, 19, 19, 19, 19, 19, 19, 19, 19, 19, 19, 19, 19, 19, 19, 19}}}, {kS8x16ZipLeft, {{0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23}}}, {kS8x16ZipRight, {{8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31}}}, {kS8x16UnzipLeft, {{0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30}}}, {kS8x16UnzipRight, {{1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31}}}, {kS8x16TransposeLeft, {{0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30}}}, {kS8x16TransposeRight, {{1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31}}}, {kS8x8Reverse, // swizzle only {{7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8}}}, {kS8x4Reverse, // swizzle only {{3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12}}}, {kS8x2Reverse, // swizzle only {{1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14}}}, {kS8x16Irregular, {{0, 16, 0, 16, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23}}}, }; #define SHUFFLE_TEST(Name) \ WASM_SIMD_TEST(Name) { \ ShuffleMap::const_iterator it = test_shuffles.find(k##Name); \ DCHECK_NE(it, test_shuffles.end()); \ RunShuffleOpTest(execution_tier, lower_simd, kExprS8x16Shuffle, \ it->second); \ } SHUFFLE_LIST(SHUFFLE_TEST) #undef SHUFFLE_TEST #undef SHUFFLE_LIST // Test shuffles that blend the two vectors (elements remain in their lanes.) WASM_SIMD_TEST(S8x16Blend) { std::array expected; for (int bias = 1; bias < kSimd128Size; bias++) { for (int i = 0; i < bias; i++) expected[i] = i; for (int i = bias; i < kSimd128Size; i++) expected[i] = i + kSimd128Size; RunShuffleOpTest(execution_tier, lower_simd, kExprS8x16Shuffle, expected); } } // Test shuffles that concatenate the two vectors. WASM_SIMD_TEST(S8x16Concat) { std::array expected; // n is offset or bias of concatenation. for (int n = 1; n < kSimd128Size; ++n) { int i = 0; // last kLanes - n bytes of first vector. for (int j = n; j < kSimd128Size; ++j) { expected[i++] = j; } // first n bytes of second vector for (int j = 0; j < n; ++j) { expected[i++] = j + kSimd128Size; } RunShuffleOpTest(execution_tier, lower_simd, kExprS8x16Shuffle, expected); } } // Combine 3 shuffles a, b, and c by applying both a and b and then applying c // to those two results. Shuffle Combine(const Shuffle& a, const Shuffle& b, const Shuffle& c) { Shuffle result; for (int i = 0; i < kSimd128Size; ++i) { result[i] = c[i] < kSimd128Size ? a[c[i]] : b[c[i] - kSimd128Size]; } return result; } const Shuffle& GetRandomTestShuffle(v8::base::RandomNumberGenerator* rng) { return test_shuffles[static_cast(rng->NextInt(kNumShuffleKeys))]; } // Test shuffles that are random combinations of 3 test shuffles. Completely // random shuffles almost always generate the slow general shuffle code, so // don't exercise as many code paths. WASM_SIMD_TEST(S8x16ShuffleFuzz) { v8::base::RandomNumberGenerator* rng = CcTest::random_number_generator(); static const int kTests = 100; for (int i = 0; i < kTests; ++i) { auto shuffle = Combine(GetRandomTestShuffle(rng), GetRandomTestShuffle(rng), GetRandomTestShuffle(rng)); RunShuffleOpTest(execution_tier, lower_simd, kExprS8x16Shuffle, shuffle); } } void AppendShuffle(const Shuffle& shuffle, std::vector* buffer) { byte opcode[] = {WASM_SIMD_OP(kExprS8x16Shuffle)}; for (size_t i = 0; i < arraysize(opcode); ++i) buffer->push_back(opcode[i]); for (size_t i = 0; i < kSimd128Size; ++i) buffer->push_back((shuffle[i])); } void BuildShuffle(std::vector& shuffles, // NOLINT(runtime/references) std::vector* buffer) { // Perform the leaf shuffles on globals 0 and 1. size_t row_index = (shuffles.size() - 1) / 2; for (size_t i = row_index; i < shuffles.size(); ++i) { byte operands[] = {WASM_GET_GLOBAL(0), WASM_GET_GLOBAL(1)}; for (size_t j = 0; j < arraysize(operands); ++j) buffer->push_back(operands[j]); AppendShuffle(shuffles[i], buffer); } // Now perform inner shuffles in the correct order on operands on the stack. do { for (size_t i = row_index / 2; i < row_index; ++i) { AppendShuffle(shuffles[i], buffer); } row_index /= 2; } while (row_index != 0); byte epilog[] = {kExprSetGlobal, static_cast(0), WASM_ONE}; for (size_t j = 0; j < arraysize(epilog); ++j) buffer->push_back(epilog[j]); } void RunWasmCode(ExecutionTier execution_tier, LowerSimd lower_simd, const std::vector& code, std::array* result) { WasmRunner r(execution_tier, lower_simd); // Set up two test patterns as globals, e.g. [0, 1, 2, 3] and [4, 5, 6, 7]. int8_t* src0 = r.builder().AddGlobal(kWasmS128); int8_t* src1 = r.builder().AddGlobal(kWasmS128); for (int i = 0; i < kSimd128Size; ++i) { WriteLittleEndianValue(&src0[i], i); WriteLittleEndianValue(&src1[i], kSimd128Size + i); } r.Build(code.data(), code.data() + code.size()); CHECK_EQ(1, r.Call()); for (size_t i = 0; i < kSimd128Size; i++) { (*result)[i] = ReadLittleEndianValue(&src0[i]); } } // Test multiple shuffles executed in sequence. WASM_SIMD_COMPILED_TEST(S8x16MultiShuffleFuzz) { v8::base::RandomNumberGenerator* rng = CcTest::random_number_generator(); static const int kShuffles = 100; for (int i = 0; i < kShuffles; ++i) { // Create an odd number in [3..23] of random test shuffles so we can build // a complete binary tree (stored as a heap) of shuffle operations. The leaf // shuffles operate on the test pattern inputs, while the interior shuffles // operate on the results of the two child shuffles. int num_shuffles = rng->NextInt(10) * 2 + 3; std::vector shuffles; for (int j = 0; j < num_shuffles; ++j) { shuffles.push_back(GetRandomTestShuffle(rng)); } // Generate the code for the shuffle expression. std::vector buffer; BuildShuffle(shuffles, &buffer); // Run the code using the interpreter to get the expected result. std::array expected; RunWasmCode(ExecutionTier::kInterpreter, kNoLowerSimd, buffer, &expected); // Run the SIMD or scalar lowered compiled code and compare results. std::array result; RunWasmCode(execution_tier, lower_simd, buffer, &result); for (size_t i = 0; i < kSimd128Size; ++i) { CHECK_EQ(result[i], expected[i]); } } } // Boolean unary operations are 'AllTrue' and 'AnyTrue', which return an integer // result. Use relational ops on numeric vectors to create the boolean vector // test inputs. Test inputs with all true, all false, one true, and one false. #define WASM_SIMD_BOOL_REDUCTION_TEST(format, lanes) \ WASM_SIMD_TEST(ReductionTest##lanes) { \ WasmRunner r(execution_tier, lower_simd); \ byte zero = r.AllocateLocal(kWasmS128); \ byte one_one = r.AllocateLocal(kWasmS128); \ byte reduced = r.AllocateLocal(kWasmI32); \ BUILD(r, WASM_SET_LOCAL(zero, WASM_SIMD_I##format##_SPLAT(WASM_ZERO)), \ WASM_SET_LOCAL( \ reduced, WASM_SIMD_UNOP(kExprS1x##lanes##AnyTrue, \ WASM_SIMD_BINOP(kExprI##format##Eq, \ WASM_GET_LOCAL(zero), \ WASM_GET_LOCAL(zero)))), \ WASM_IF(WASM_I32_EQ(WASM_GET_LOCAL(reduced), WASM_ZERO), \ WASM_RETURN1(WASM_ZERO)), \ WASM_SET_LOCAL( \ reduced, WASM_SIMD_UNOP(kExprS1x##lanes##AnyTrue, \ WASM_SIMD_BINOP(kExprI##format##Ne, \ WASM_GET_LOCAL(zero), \ WASM_GET_LOCAL(zero)))), \ WASM_IF(WASM_I32_NE(WASM_GET_LOCAL(reduced), WASM_ZERO), \ WASM_RETURN1(WASM_ZERO)), \ WASM_SET_LOCAL( \ reduced, WASM_SIMD_UNOP(kExprS1x##lanes##AllTrue, \ WASM_SIMD_BINOP(kExprI##format##Eq, \ WASM_GET_LOCAL(zero), \ WASM_GET_LOCAL(zero)))), \ WASM_IF(WASM_I32_EQ(WASM_GET_LOCAL(reduced), WASM_ZERO), \ WASM_RETURN1(WASM_ZERO)), \ WASM_SET_LOCAL( \ reduced, WASM_SIMD_UNOP(kExprS1x##lanes##AllTrue, \ WASM_SIMD_BINOP(kExprI##format##Ne, \ WASM_GET_LOCAL(zero), \ WASM_GET_LOCAL(zero)))), \ WASM_IF(WASM_I32_NE(WASM_GET_LOCAL(reduced), WASM_ZERO), \ WASM_RETURN1(WASM_ZERO)), \ WASM_SET_LOCAL(one_one, \ WASM_SIMD_I##format##_REPLACE_LANE( \ lanes - 1, WASM_GET_LOCAL(zero), WASM_ONE)), \ WASM_SET_LOCAL( \ reduced, WASM_SIMD_UNOP(kExprS1x##lanes##AnyTrue, \ WASM_SIMD_BINOP(kExprI##format##Eq, \ WASM_GET_LOCAL(one_one), \ WASM_GET_LOCAL(zero)))), \ WASM_IF(WASM_I32_EQ(WASM_GET_LOCAL(reduced), WASM_ZERO), \ WASM_RETURN1(WASM_ZERO)), \ WASM_SET_LOCAL( \ reduced, WASM_SIMD_UNOP(kExprS1x##lanes##AnyTrue, \ WASM_SIMD_BINOP(kExprI##format##Ne, \ WASM_GET_LOCAL(one_one), \ WASM_GET_LOCAL(zero)))), \ WASM_IF(WASM_I32_EQ(WASM_GET_LOCAL(reduced), WASM_ZERO), \ WASM_RETURN1(WASM_ZERO)), \ WASM_SET_LOCAL( \ reduced, WASM_SIMD_UNOP(kExprS1x##lanes##AllTrue, \ WASM_SIMD_BINOP(kExprI##format##Eq, \ WASM_GET_LOCAL(one_one), \ WASM_GET_LOCAL(zero)))), \ WASM_IF(WASM_I32_NE(WASM_GET_LOCAL(reduced), WASM_ZERO), \ WASM_RETURN1(WASM_ZERO)), \ WASM_SET_LOCAL( \ reduced, WASM_SIMD_UNOP(kExprS1x##lanes##AllTrue, \ WASM_SIMD_BINOP(kExprI##format##Ne, \ WASM_GET_LOCAL(one_one), \ WASM_GET_LOCAL(zero)))), \ WASM_IF(WASM_I32_NE(WASM_GET_LOCAL(reduced), WASM_ZERO), \ WASM_RETURN1(WASM_ZERO)), \ WASM_ONE); \ CHECK_EQ(1, r.Call()); \ } WASM_SIMD_BOOL_REDUCTION_TEST(32x4, 4) WASM_SIMD_BOOL_REDUCTION_TEST(16x8, 8) WASM_SIMD_BOOL_REDUCTION_TEST(8x16, 16) WASM_SIMD_TEST(SimdI32x4ExtractWithF32x4) { WasmRunner r(execution_tier, lower_simd); BUILD(r, WASM_IF_ELSE_I( WASM_I32_EQ(WASM_SIMD_I32x4_EXTRACT_LANE( 0, WASM_SIMD_F32x4_SPLAT(WASM_F32(30.5))), WASM_I32_REINTERPRET_F32(WASM_F32(30.5))), WASM_I32V(1), WASM_I32V(0))); CHECK_EQ(1, r.Call()); } WASM_SIMD_TEST(SimdF32x4ExtractWithI32x4) { WasmRunner r(execution_tier, lower_simd); BUILD(r, WASM_IF_ELSE_I(WASM_F32_EQ(WASM_SIMD_F32x4_EXTRACT_LANE( 0, WASM_SIMD_I32x4_SPLAT(WASM_I32V(15))), WASM_F32_REINTERPRET_I32(WASM_I32V(15))), WASM_I32V(1), WASM_I32V(0))); CHECK_EQ(1, r.Call()); } WASM_SIMD_TEST(SimdF32x4ExtractLane) { WasmRunner r(execution_tier, lower_simd); r.AllocateLocal(kWasmF32); r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(0, WASM_SIMD_F32x4_EXTRACT_LANE( 0, WASM_SIMD_F32x4_SPLAT(WASM_F32(30.5)))), WASM_SET_LOCAL(1, WASM_SIMD_F32x4_SPLAT(WASM_GET_LOCAL(0))), WASM_SIMD_F32x4_EXTRACT_LANE(1, WASM_GET_LOCAL(1))); CHECK_EQ(30.5, r.Call()); } WASM_SIMD_TEST(SimdF32x4AddWithI32x4) { // Choose two floating point values whose sum is normal and exactly // representable as a float. const int kOne = 0x3F800000; const int kTwo = 0x40000000; WasmRunner r(execution_tier, lower_simd); BUILD(r, WASM_IF_ELSE_I( WASM_F32_EQ( WASM_SIMD_F32x4_EXTRACT_LANE( 0, WASM_SIMD_BINOP(kExprF32x4Add, WASM_SIMD_I32x4_SPLAT(WASM_I32V(kOne)), WASM_SIMD_I32x4_SPLAT(WASM_I32V(kTwo)))), WASM_F32_ADD(WASM_F32_REINTERPRET_I32(WASM_I32V(kOne)), WASM_F32_REINTERPRET_I32(WASM_I32V(kTwo)))), WASM_I32V(1), WASM_I32V(0))); CHECK_EQ(1, r.Call()); } WASM_SIMD_TEST(SimdI32x4AddWithF32x4) { WasmRunner r(execution_tier, lower_simd); BUILD(r, WASM_IF_ELSE_I( WASM_I32_EQ( WASM_SIMD_I32x4_EXTRACT_LANE( 0, WASM_SIMD_BINOP(kExprI32x4Add, WASM_SIMD_F32x4_SPLAT(WASM_F32(21.25)), WASM_SIMD_F32x4_SPLAT(WASM_F32(31.5)))), WASM_I32_ADD(WASM_I32_REINTERPRET_F32(WASM_F32(21.25)), WASM_I32_REINTERPRET_F32(WASM_F32(31.5)))), WASM_I32V(1), WASM_I32V(0))); CHECK_EQ(1, r.Call()); } WASM_SIMD_TEST(SimdI32x4Local) { WasmRunner r(execution_tier, lower_simd); r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(0, WASM_SIMD_I32x4_SPLAT(WASM_I32V(31))), WASM_SIMD_I32x4_EXTRACT_LANE(0, WASM_GET_LOCAL(0))); CHECK_EQ(31, r.Call()); } WASM_SIMD_TEST(SimdI32x4SplatFromExtract) { WasmRunner r(execution_tier, lower_simd); r.AllocateLocal(kWasmI32); r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(0, WASM_SIMD_I32x4_EXTRACT_LANE( 0, WASM_SIMD_I32x4_SPLAT(WASM_I32V(76)))), WASM_SET_LOCAL(1, WASM_SIMD_I32x4_SPLAT(WASM_GET_LOCAL(0))), WASM_SIMD_I32x4_EXTRACT_LANE(1, WASM_GET_LOCAL(1))); CHECK_EQ(76, r.Call()); } WASM_SIMD_TEST(SimdI32x4For) { WasmRunner r(execution_tier, lower_simd); r.AllocateLocal(kWasmI32); r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(1, WASM_SIMD_I32x4_SPLAT(WASM_I32V(31))), WASM_SET_LOCAL(1, WASM_SIMD_I32x4_REPLACE_LANE(1, WASM_GET_LOCAL(1), WASM_I32V(53))), WASM_SET_LOCAL(1, WASM_SIMD_I32x4_REPLACE_LANE(2, WASM_GET_LOCAL(1), WASM_I32V(23))), WASM_SET_LOCAL(0, WASM_I32V(0)), WASM_LOOP( WASM_SET_LOCAL( 1, WASM_SIMD_BINOP(kExprI32x4Add, WASM_GET_LOCAL(1), WASM_SIMD_I32x4_SPLAT(WASM_I32V(1)))), WASM_IF(WASM_I32_NE(WASM_INC_LOCAL(0), WASM_I32V(5)), WASM_BR(1))), WASM_SET_LOCAL(0, WASM_I32V(1)), WASM_IF(WASM_I32_NE(WASM_SIMD_I32x4_EXTRACT_LANE(0, WASM_GET_LOCAL(1)), WASM_I32V(36)), WASM_SET_LOCAL(0, WASM_I32V(0))), WASM_IF(WASM_I32_NE(WASM_SIMD_I32x4_EXTRACT_LANE(1, WASM_GET_LOCAL(1)), WASM_I32V(58)), WASM_SET_LOCAL(0, WASM_I32V(0))), WASM_IF(WASM_I32_NE(WASM_SIMD_I32x4_EXTRACT_LANE(2, WASM_GET_LOCAL(1)), WASM_I32V(28)), WASM_SET_LOCAL(0, WASM_I32V(0))), WASM_IF(WASM_I32_NE(WASM_SIMD_I32x4_EXTRACT_LANE(3, WASM_GET_LOCAL(1)), WASM_I32V(36)), WASM_SET_LOCAL(0, WASM_I32V(0))), WASM_GET_LOCAL(0)); CHECK_EQ(1, r.Call()); } WASM_SIMD_TEST(SimdF32x4For) { WasmRunner r(execution_tier, lower_simd); r.AllocateLocal(kWasmI32); r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(1, WASM_SIMD_F32x4_SPLAT(WASM_F32(21.25))), WASM_SET_LOCAL(1, WASM_SIMD_F32x4_REPLACE_LANE(3, WASM_GET_LOCAL(1), WASM_F32(19.5))), WASM_SET_LOCAL(0, WASM_I32V(0)), WASM_LOOP( WASM_SET_LOCAL( 1, WASM_SIMD_BINOP(kExprF32x4Add, WASM_GET_LOCAL(1), WASM_SIMD_F32x4_SPLAT(WASM_F32(2.0)))), WASM_IF(WASM_I32_NE(WASM_INC_LOCAL(0), WASM_I32V(3)), WASM_BR(1))), WASM_SET_LOCAL(0, WASM_I32V(1)), WASM_IF(WASM_F32_NE(WASM_SIMD_F32x4_EXTRACT_LANE(0, WASM_GET_LOCAL(1)), WASM_F32(27.25)), WASM_SET_LOCAL(0, WASM_I32V(0))), WASM_IF(WASM_F32_NE(WASM_SIMD_F32x4_EXTRACT_LANE(3, WASM_GET_LOCAL(1)), WASM_F32(25.5)), WASM_SET_LOCAL(0, WASM_I32V(0))), WASM_GET_LOCAL(0)); CHECK_EQ(1, r.Call()); } template void SetVectorByLanes(T* v, const std::array& arr) { for (int lane = 0; lane < numLanes; lane++) { WriteLittleEndianValue(&v[lane], arr[lane]); } } template const T GetScalar(T* v, int lane) { constexpr int kElems = kSimd128Size / sizeof(T); const int index = lane; USE(kElems); DCHECK(index >= 0 && index < kElems); return ReadLittleEndianValue(&v[index]); } WASM_SIMD_TEST(SimdI32x4GetGlobal) { WasmRunner r(execution_tier, lower_simd); // Pad the globals with a few unused slots to get a non-zero offset. r.builder().AddGlobal(kWasmI32); // purposefully unused r.builder().AddGlobal(kWasmI32); // purposefully unused r.builder().AddGlobal(kWasmI32); // purposefully unused r.builder().AddGlobal(kWasmI32); // purposefully unused int32_t* global = r.builder().AddGlobal(kWasmS128); SetVectorByLanes(global, {{0, 1, 2, 3}}); r.AllocateLocal(kWasmI32); BUILD( r, WASM_SET_LOCAL(1, WASM_I32V(1)), WASM_IF(WASM_I32_NE(WASM_I32V(0), WASM_SIMD_I32x4_EXTRACT_LANE(0, WASM_GET_GLOBAL(4))), WASM_SET_LOCAL(1, WASM_I32V(0))), WASM_IF(WASM_I32_NE(WASM_I32V(1), WASM_SIMD_I32x4_EXTRACT_LANE(1, WASM_GET_GLOBAL(4))), WASM_SET_LOCAL(1, WASM_I32V(0))), WASM_IF(WASM_I32_NE(WASM_I32V(2), WASM_SIMD_I32x4_EXTRACT_LANE(2, WASM_GET_GLOBAL(4))), WASM_SET_LOCAL(1, WASM_I32V(0))), WASM_IF(WASM_I32_NE(WASM_I32V(3), WASM_SIMD_I32x4_EXTRACT_LANE(3, WASM_GET_GLOBAL(4))), WASM_SET_LOCAL(1, WASM_I32V(0))), WASM_GET_LOCAL(1)); CHECK_EQ(1, r.Call(0)); } WASM_SIMD_TEST(SimdI32x4SetGlobal) { WasmRunner r(execution_tier, lower_simd); // Pad the globals with a few unused slots to get a non-zero offset. r.builder().AddGlobal(kWasmI32); // purposefully unused r.builder().AddGlobal(kWasmI32); // purposefully unused r.builder().AddGlobal(kWasmI32); // purposefully unused r.builder().AddGlobal(kWasmI32); // purposefully unused int32_t* global = r.builder().AddGlobal(kWasmS128); BUILD(r, WASM_SET_GLOBAL(4, WASM_SIMD_I32x4_SPLAT(WASM_I32V(23))), WASM_SET_GLOBAL(4, WASM_SIMD_I32x4_REPLACE_LANE(1, WASM_GET_GLOBAL(4), WASM_I32V(34))), WASM_SET_GLOBAL(4, WASM_SIMD_I32x4_REPLACE_LANE(2, WASM_GET_GLOBAL(4), WASM_I32V(45))), WASM_SET_GLOBAL(4, WASM_SIMD_I32x4_REPLACE_LANE(3, WASM_GET_GLOBAL(4), WASM_I32V(56))), WASM_I32V(1)); CHECK_EQ(1, r.Call(0)); CHECK_EQ(GetScalar(global, 0), 23); CHECK_EQ(GetScalar(global, 1), 34); CHECK_EQ(GetScalar(global, 2), 45); CHECK_EQ(GetScalar(global, 3), 56); } WASM_SIMD_TEST(SimdF32x4GetGlobal) { WasmRunner r(execution_tier, lower_simd); float* global = r.builder().AddGlobal(kWasmS128); SetVectorByLanes(global, {{0.0, 1.5, 2.25, 3.5}}); r.AllocateLocal(kWasmI32); BUILD( r, WASM_SET_LOCAL(1, WASM_I32V(1)), WASM_IF(WASM_F32_NE(WASM_F32(0.0), WASM_SIMD_F32x4_EXTRACT_LANE(0, WASM_GET_GLOBAL(0))), WASM_SET_LOCAL(1, WASM_I32V(0))), WASM_IF(WASM_F32_NE(WASM_F32(1.5), WASM_SIMD_F32x4_EXTRACT_LANE(1, WASM_GET_GLOBAL(0))), WASM_SET_LOCAL(1, WASM_I32V(0))), WASM_IF(WASM_F32_NE(WASM_F32(2.25), WASM_SIMD_F32x4_EXTRACT_LANE(2, WASM_GET_GLOBAL(0))), WASM_SET_LOCAL(1, WASM_I32V(0))), WASM_IF(WASM_F32_NE(WASM_F32(3.5), WASM_SIMD_F32x4_EXTRACT_LANE(3, WASM_GET_GLOBAL(0))), WASM_SET_LOCAL(1, WASM_I32V(0))), WASM_GET_LOCAL(1)); CHECK_EQ(1, r.Call(0)); } WASM_SIMD_TEST(SimdF32x4SetGlobal) { WasmRunner r(execution_tier, lower_simd); float* global = r.builder().AddGlobal(kWasmS128); BUILD(r, WASM_SET_GLOBAL(0, WASM_SIMD_F32x4_SPLAT(WASM_F32(13.5))), WASM_SET_GLOBAL(0, WASM_SIMD_F32x4_REPLACE_LANE(1, WASM_GET_GLOBAL(0), WASM_F32(45.5))), WASM_SET_GLOBAL(0, WASM_SIMD_F32x4_REPLACE_LANE(2, WASM_GET_GLOBAL(0), WASM_F32(32.25))), WASM_SET_GLOBAL(0, WASM_SIMD_F32x4_REPLACE_LANE(3, WASM_GET_GLOBAL(0), WASM_F32(65.0))), WASM_I32V(1)); CHECK_EQ(1, r.Call(0)); CHECK_EQ(GetScalar(global, 0), 13.5f); CHECK_EQ(GetScalar(global, 1), 45.5f); CHECK_EQ(GetScalar(global, 2), 32.25f); CHECK_EQ(GetScalar(global, 3), 65.0f); } WASM_SIMD_COMPILED_TEST(SimdLoadStoreLoad) { WasmRunner r(execution_tier, lower_simd); int32_t* memory = r.builder().AddMemoryElems(kWasmPageSize / sizeof(int32_t)); // Load memory, store it, then reload it and extract the first lane. Use a // non-zero offset into the memory of 1 lane (4 bytes) to test indexing. BUILD(r, WASM_SIMD_STORE_MEM(WASM_I32V(4), WASM_SIMD_LOAD_MEM(WASM_I32V(4))), WASM_SIMD_I32x4_EXTRACT_LANE(0, WASM_SIMD_LOAD_MEM(WASM_I32V(4)))); FOR_INT32_INPUTS(i) { int32_t expected = i; r.builder().WriteMemory(&memory[1], expected); CHECK_EQ(expected, r.Call()); } } #if V8_TARGET_ARCH_X64 || V8_TARGET_ARCH_IA32 || V8_TARGET_ARCH_ARM64 || \ V8_TARGET_ARCH_ARM // V8:8665 - Tracking bug to enable reduction tests in the interpreter, // and for SIMD lowering. #define WASM_SIMD_ANYTRUE_TEST(format, lanes, max, param_type) \ WASM_SIMD_TEST_NO_LOWERING(S##format##AnyTrue) { \ WasmRunner r(execution_tier, lower_simd); \ byte simd = r.AllocateLocal(kWasmS128); \ BUILD( \ r, \ WASM_SET_LOCAL(simd, WASM_SIMD_I##format##_SPLAT(WASM_GET_LOCAL(0))), \ WASM_SIMD_UNOP(kExprS1x##lanes##AnyTrue, WASM_GET_LOCAL(simd))); \ DCHECK_EQ(1, r.Call(max)); \ DCHECK_EQ(1, r.Call(5)); \ DCHECK_EQ(0, r.Call(0)); \ } #if V8_TARGET_ARCH_X64 WASM_SIMD_ANYTRUE_TEST(64x2, 2, 0xffffffffffffffff, int64_t) #endif // V8_TARGET_ARCH_X64 WASM_SIMD_ANYTRUE_TEST(32x4, 4, 0xffffffff, int32_t) WASM_SIMD_ANYTRUE_TEST(16x8, 8, 0xffff, int32_t) WASM_SIMD_ANYTRUE_TEST(8x16, 16, 0xff, int32_t) #define WASM_SIMD_ALLTRUE_TEST(format, lanes, max, param_type) \ WASM_SIMD_TEST_NO_LOWERING(S##format##AllTrue) { \ WasmRunner r(execution_tier, lower_simd); \ byte simd = r.AllocateLocal(kWasmS128); \ BUILD( \ r, \ WASM_SET_LOCAL(simd, WASM_SIMD_I##format##_SPLAT(WASM_GET_LOCAL(0))), \ WASM_SIMD_UNOP(kExprS1x##lanes##AllTrue, WASM_GET_LOCAL(simd))); \ DCHECK_EQ(1, r.Call(max)); \ DCHECK_EQ(1, r.Call(0x1)); \ DCHECK_EQ(0, r.Call(0)); \ } #if V8_TARGET_ARCH_X64 WASM_SIMD_ALLTRUE_TEST(64x2, 2, 0xffffffffffffffff, int64_t) #endif // V8_TARGET_ARCH_X64 WASM_SIMD_ALLTRUE_TEST(32x4, 4, 0xffffffff, int32_t) WASM_SIMD_ALLTRUE_TEST(16x8, 8, 0xffff, int32_t) WASM_SIMD_ALLTRUE_TEST(8x16, 16, 0xff, int32_t) #endif // V8_TARGET_ARCH_X64 || V8_TARGET_ARCH_IA32 || V8_TARGET_ARCH_ARM64 || // V8_TARGET_ARCH_ARM WASM_SIMD_TEST(BitSelect) { WasmRunner r(execution_tier, lower_simd); byte simd = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL( simd, WASM_SIMD_SELECT(32x4, WASM_SIMD_I32x4_SPLAT(WASM_I32V(0x01020304)), WASM_SIMD_I32x4_SPLAT(WASM_I32V(0)), WASM_SIMD_I32x4_SPLAT(WASM_GET_LOCAL(0)))), WASM_SIMD_I32x4_EXTRACT_LANE(0, WASM_GET_LOCAL(simd))); DCHECK_EQ(0x01020304, r.Call(0xFFFFFFFF)); } void RunI8x16MixedRelationalOpTest(ExecutionTier execution_tier, LowerSimd lower_simd, WasmOpcode opcode, Int8BinOp expected_op) { WasmRunner r(execution_tier, lower_simd); byte value1 = 0, value2 = 1; byte temp1 = r.AllocateLocal(kWasmS128); byte temp2 = r.AllocateLocal(kWasmS128); byte temp3 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_I8x16_SPLAT(WASM_GET_LOCAL(value1))), WASM_SET_LOCAL(temp2, WASM_SIMD_I16x8_SPLAT(WASM_GET_LOCAL(value2))), WASM_SET_LOCAL(temp3, WASM_SIMD_BINOP(opcode, WASM_GET_LOCAL(temp1), WASM_GET_LOCAL(temp2))), WASM_SIMD_I8x16_EXTRACT_LANE(0, WASM_GET_LOCAL(temp3))); DCHECK_EQ(expected_op(0xff, static_cast(0x7fff)), r.Call(0xff, 0x7fff)); DCHECK_EQ(expected_op(0xfe, static_cast(0x7fff)), r.Call(0xfe, 0x7fff)); DCHECK_EQ(expected_op(0xff, static_cast(0x7ffe)), r.Call(0xff, 0x7ffe)); } WASM_SIMD_TEST_NO_LOWERING(I8x16LeUMixed) { RunI8x16MixedRelationalOpTest(execution_tier, lower_simd, kExprI8x16LeU, UnsignedLessEqual); } WASM_SIMD_TEST_NO_LOWERING(I8x16LtUMixed) { RunI8x16MixedRelationalOpTest(execution_tier, lower_simd, kExprI8x16LtU, UnsignedLess); } WASM_SIMD_TEST_NO_LOWERING(I8x16GeUMixed) { RunI8x16MixedRelationalOpTest(execution_tier, lower_simd, kExprI8x16GeU, UnsignedGreaterEqual); } WASM_SIMD_TEST_NO_LOWERING(I8x16GtUMixed) { RunI8x16MixedRelationalOpTest(execution_tier, lower_simd, kExprI8x16GtU, UnsignedGreater); } void RunI16x8MixedRelationalOpTest(ExecutionTier execution_tier, LowerSimd lower_simd, WasmOpcode opcode, Int16BinOp expected_op) { WasmRunner r(execution_tier, lower_simd); byte value1 = 0, value2 = 1; byte temp1 = r.AllocateLocal(kWasmS128); byte temp2 = r.AllocateLocal(kWasmS128); byte temp3 = r.AllocateLocal(kWasmS128); BUILD(r, WASM_SET_LOCAL(temp1, WASM_SIMD_I16x8_SPLAT(WASM_GET_LOCAL(value1))), WASM_SET_LOCAL(temp2, WASM_SIMD_I32x4_SPLAT(WASM_GET_LOCAL(value2))), WASM_SET_LOCAL(temp3, WASM_SIMD_BINOP(opcode, WASM_GET_LOCAL(temp1), WASM_GET_LOCAL(temp2))), WASM_SIMD_I16x8_EXTRACT_LANE(0, WASM_GET_LOCAL(temp3))); DCHECK_EQ(expected_op(0xffff, static_cast(0x7fffffff)), r.Call(0xffff, 0x7fffffff)); DCHECK_EQ(expected_op(0xfeff, static_cast(0x7fffffff)), r.Call(0xfeff, 0x7fffffff)); DCHECK_EQ(expected_op(0xffff, static_cast(0x7ffffeff)), r.Call(0xffff, 0x7ffffeff)); } WASM_SIMD_TEST_NO_LOWERING(I16x8LeUMixed) { RunI16x8MixedRelationalOpTest(execution_tier, lower_simd, kExprI16x8LeU, UnsignedLessEqual); } WASM_SIMD_TEST_NO_LOWERING(I16x8LtUMixed) { RunI16x8MixedRelationalOpTest(execution_tier, lower_simd, kExprI16x8LtU, UnsignedLess); } WASM_SIMD_TEST_NO_LOWERING(I16x8GeUMixed) { RunI16x8MixedRelationalOpTest(execution_tier, lower_simd, kExprI16x8GeU, UnsignedGreaterEqual); } WASM_SIMD_TEST_NO_LOWERING(I16x8GtUMixed) { RunI16x8MixedRelationalOpTest(execution_tier, lower_simd, kExprI16x8GtU, UnsignedGreater); } #undef WASM_SIMD_TEST #undef WASM_SIMD_CHECK_LANE #undef TO_BYTE #undef WASM_SIMD_OP #undef WASM_SIMD_SPLAT #undef WASM_SIMD_UNOP #undef WASM_SIMD_BINOP #undef WASM_SIMD_SHIFT_OP #undef WASM_SIMD_CONCAT_OP #undef WASM_SIMD_SELECT #undef WASM_SIMD_F64x2_SPLAT #undef WASM_SIMD_F64x2_EXTRACT_LANE #undef WASM_SIMD_F64x2_REPLACE_LANE #undef WASM_SIMD_F32x4_SPLAT #undef WASM_SIMD_F32x4_EXTRACT_LANE #undef WASM_SIMD_F32x4_REPLACE_LANE #undef WASM_SIMD_I64x2_SPLAT #undef WASM_SIMD_I64x2_EXTRACT_LANE #undef WASM_SIMD_I64x2_REPLACE_LANE #undef WASM_SIMD_I32x4_SPLAT #undef WASM_SIMD_I32x4_EXTRACT_LANE #undef WASM_SIMD_I32x4_REPLACE_LANE #undef WASM_SIMD_I16x8_SPLAT #undef WASM_SIMD_I16x8_EXTRACT_LANE #undef WASM_SIMD_I16x8_REPLACE_LANE #undef WASM_SIMD_I8x16_SPLAT #undef WASM_SIMD_I8x16_EXTRACT_LANE #undef WASM_SIMD_I8x16_REPLACE_LANE #undef WASM_SIMD_S8x16_SHUFFLE_OP #undef WASM_SIMD_LOAD_MEM #undef WASM_SIMD_STORE_MEM #undef WASM_SIMD_SELECT_TEST #undef WASM_SIMD_NON_CANONICAL_SELECT_TEST #undef WASM_SIMD_COMPILED_TEST #undef WASM_SIMD_BOOL_REDUCTION_TEST #undef WASM_SIMD_TEST_NO_LOWERING #undef WASM_SIMD_ANYTRUE_TEST #undef WASM_SIMD_ALLTRUE_TEST } // namespace test_run_wasm_simd } // namespace wasm } // namespace internal } // namespace v8