7260f62b88
This CL * Adds the xadd instruction to the ia32 assembler and disassembler; * Implements the AtomicAdd instructions, except AtomicAddU64, on ia32; R=clemensb@chromium.org Bug: v8:10108 Change-Id: Ic8653a9f96148282951104fefb4185c4c0db89a3 Reviewed-on: https://chromium-review.googlesource.com/c/v8/v8/+/2232719 Reviewed-by: Clemens Backes <clemensb@chromium.org> Commit-Queue: Andreas Haas <ahaas@chromium.org> Cr-Commit-Position: refs/heads/master@{#68247}
1003 lines
33 KiB
C++
1003 lines
33 KiB
C++
// Copyright 2011 the V8 project authors. All rights reserved.
|
|
// Redistribution and use in source and binary forms, with or without
|
|
// modification, are permitted provided that the following conditions are
|
|
// met:
|
|
//
|
|
// * Redistributions of source code must retain the above copyright
|
|
// notice, this list of conditions and the following disclaimer.
|
|
// * Redistributions in binary form must reproduce the above
|
|
// copyright notice, this list of conditions and the following
|
|
// disclaimer in the documentation and/or other materials provided
|
|
// with the distribution.
|
|
// * Neither the name of Google Inc. nor the names of its
|
|
// contributors may be used to endorse or promote products derived
|
|
// from this software without specific prior written permission.
|
|
//
|
|
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
#include <stdlib.h>
|
|
|
|
#include "src/init/v8.h"
|
|
|
|
#include "src/codegen/code-factory.h"
|
|
#include "src/codegen/macro-assembler.h"
|
|
#include "src/debug/debug.h"
|
|
#include "src/diagnostics/disasm.h"
|
|
#include "src/diagnostics/disassembler.h"
|
|
#include "src/execution/frames-inl.h"
|
|
#include "src/utils/ostreams.h"
|
|
#include "test/cctest/cctest.h"
|
|
|
|
namespace v8 {
|
|
namespace internal {
|
|
|
|
#define __ assm.
|
|
|
|
static void DummyStaticFunction(Object result) {}
|
|
|
|
TEST(DisasmIa320) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
v8::internal::byte buffer[8192];
|
|
Assembler assm(AssemblerOptions{},
|
|
ExternalAssemblerBuffer(buffer, sizeof buffer));
|
|
// Short immediate instructions
|
|
__ adc(eax, 12345678);
|
|
__ add(eax, Immediate(12345678));
|
|
__ or_(eax, 12345678);
|
|
__ sub(eax, Immediate(12345678));
|
|
__ xor_(eax, 12345678);
|
|
__ and_(eax, 12345678);
|
|
Handle<FixedArray> foo =
|
|
isolate->factory()->NewFixedArray(10, AllocationType::kOld);
|
|
__ cmp(eax, foo);
|
|
|
|
// ---- This one caused crash
|
|
__ mov(ebx, Operand(esp, ecx, times_2, 0)); // [esp+ecx*4]
|
|
|
|
// ---- All instructions that I can think of
|
|
__ add(edx, ebx);
|
|
__ add(edx, Operand(12, RelocInfo::NONE));
|
|
__ add(edx, Operand(ebx, 0));
|
|
__ add(edx, Operand(ebx, 16));
|
|
__ add(edx, Operand(ebx, 1999));
|
|
__ add(edx, Operand(ebx, -4));
|
|
__ add(edx, Operand(ebx, -1999));
|
|
__ add(edx, Operand(esp, 0));
|
|
__ add(edx, Operand(esp, 16));
|
|
__ add(edx, Operand(esp, 1999));
|
|
__ add(edx, Operand(esp, -4));
|
|
__ add(edx, Operand(esp, -1999));
|
|
__ nop();
|
|
__ add(esi, Operand(ecx, times_4, 0));
|
|
__ add(esi, Operand(ecx, times_4, 24));
|
|
__ add(esi, Operand(ecx, times_4, -4));
|
|
__ add(esi, Operand(ecx, times_4, -1999));
|
|
__ nop();
|
|
__ add(edi, Operand(ebp, ecx, times_4, 0));
|
|
__ add(edi, Operand(ebp, ecx, times_4, 12));
|
|
__ add(edi, Operand(ebp, ecx, times_4, -8));
|
|
__ add(edi, Operand(ebp, ecx, times_4, -3999));
|
|
__ add(Operand(ebp, ecx, times_4, 12), Immediate(12));
|
|
|
|
__ bswap(eax);
|
|
|
|
__ nop();
|
|
__ add(ebx, Immediate(12));
|
|
__ nop();
|
|
__ adc(edx, Operand(ebx));
|
|
__ adc(ecx, 12);
|
|
__ adc(ecx, 1000);
|
|
__ nop();
|
|
__ and_(edx, 3);
|
|
__ and_(edx, Operand(esp, 4));
|
|
__ cmp(edx, 3);
|
|
__ cmp(edx, Operand(esp, 4));
|
|
__ cmp(Operand(ebp, ecx, times_4, 0), Immediate(1000));
|
|
Handle<FixedArray> foo2 =
|
|
isolate->factory()->NewFixedArray(10, AllocationType::kOld);
|
|
__ cmp(ebx, foo2);
|
|
__ cmpb(ebx, Operand(ebp, ecx, times_2, 0));
|
|
__ cmpb(Operand(ebp, ecx, times_2, 0), ebx);
|
|
__ or_(edx, 3);
|
|
__ xor_(edx, 3);
|
|
__ nop();
|
|
__ cpuid();
|
|
__ movsx_b(edx, ecx);
|
|
__ movsx_w(edx, ecx);
|
|
__ movzx_b(edx, ecx);
|
|
__ movzx_w(edx, ecx);
|
|
|
|
__ nop();
|
|
__ imul(edx, ecx);
|
|
__ shld(edx, ecx, 10);
|
|
__ shld_cl(edx, ecx);
|
|
__ shrd(edx, ecx, 10);
|
|
__ shrd_cl(edx, ecx);
|
|
__ bts(edx, ecx);
|
|
__ bts(Operand(ebx, ecx, times_4, 0), ecx);
|
|
__ nop();
|
|
__ pushad();
|
|
__ popad();
|
|
__ pushfd();
|
|
__ popfd();
|
|
__ push(Immediate(12));
|
|
__ push(Immediate(23456));
|
|
__ push(ecx);
|
|
__ push(esi);
|
|
__ push(Operand(ebp, StandardFrameConstants::kFunctionOffset));
|
|
__ push(Operand(ebx, ecx, times_4, 0));
|
|
__ push(Operand(ebx, ecx, times_4, 0));
|
|
__ push(Operand(ebx, ecx, times_4, 10000));
|
|
__ pop(edx);
|
|
__ pop(eax);
|
|
__ pop(Operand(ebx, ecx, times_4, 0));
|
|
__ nop();
|
|
|
|
__ add(edx, Operand(esp, 16));
|
|
__ add(edx, ecx);
|
|
__ mov_b(edx, ecx);
|
|
__ mov_b(ecx, 6);
|
|
__ mov_b(Operand(ebx, ecx, times_4, 10000), 6);
|
|
__ mov_b(Operand(esp, 16), edx);
|
|
__ mov_w(edx, Operand(esp, 16));
|
|
__ mov_w(Operand(esp, 16), edx);
|
|
__ nop();
|
|
__ movsx_w(edx, Operand(esp, 12));
|
|
__ movsx_b(edx, Operand(esp, 12));
|
|
__ movzx_w(edx, Operand(esp, 12));
|
|
__ movzx_b(edx, Operand(esp, 12));
|
|
__ nop();
|
|
__ mov(edx, 1234567);
|
|
__ mov(edx, Operand(esp, 12));
|
|
__ mov(Operand(ebx, ecx, times_4, 10000), Immediate(12345));
|
|
__ mov(Operand(ebx, ecx, times_4, 10000), edx);
|
|
__ nop();
|
|
__ dec_b(edx);
|
|
__ dec_b(Operand(eax, 10));
|
|
__ dec_b(Operand(ebx, ecx, times_4, 10000));
|
|
__ dec(edx);
|
|
__ cdq();
|
|
|
|
__ nop();
|
|
__ idiv(edx);
|
|
__ idiv(Operand(edx, ecx, times_1, 1));
|
|
__ idiv(Operand(esp, 12));
|
|
__ div(edx);
|
|
__ div(Operand(edx, ecx, times_1, 1));
|
|
__ div(Operand(esp, 12));
|
|
__ mul(edx);
|
|
__ neg(edx);
|
|
__ not_(edx);
|
|
__ test(Operand(ebx, ecx, times_4, 10000), Immediate(123456));
|
|
|
|
__ imul(edx, Operand(ebx, ecx, times_4, 10000));
|
|
__ imul(edx, ecx, 12);
|
|
__ imul(edx, Operand(edx, eax, times_2, 42), 8);
|
|
__ imul(edx, ecx, 1000);
|
|
__ imul(edx, Operand(ebx, ecx, times_4, 1), 9000);
|
|
|
|
__ inc(edx);
|
|
__ inc(Operand(ebx, ecx, times_4, 10000));
|
|
__ push(Operand(ebx, ecx, times_4, 10000));
|
|
__ pop(Operand(ebx, ecx, times_4, 10000));
|
|
__ call(Operand(ebx, ecx, times_4, 10000));
|
|
__ jmp(Operand(ebx, ecx, times_4, 10000));
|
|
|
|
__ lea(edx, Operand(ebx, ecx, times_4, 10000));
|
|
__ or_(edx, 12345);
|
|
__ or_(edx, Operand(ebx, ecx, times_4, 10000));
|
|
|
|
__ nop();
|
|
|
|
__ rcl(edx, 1);
|
|
__ rcl(edx, 7);
|
|
__ rcr(edx, 1);
|
|
__ rcr(edx, 7);
|
|
__ ror(edx, 1);
|
|
__ ror(edx, 6);
|
|
__ ror_cl(edx);
|
|
__ ror(Operand(ebx, ecx, times_4, 10000), 1);
|
|
__ ror(Operand(ebx, ecx, times_4, 10000), 6);
|
|
__ ror_cl(Operand(ebx, ecx, times_4, 10000));
|
|
__ sar(edx, 1);
|
|
__ sar(edx, 6);
|
|
__ sar_cl(edx);
|
|
__ sar(Operand(ebx, ecx, times_4, 10000), 1);
|
|
__ sar(Operand(ebx, ecx, times_4, 10000), 6);
|
|
__ sar_cl(Operand(ebx, ecx, times_4, 10000));
|
|
__ sbb(edx, Operand(ebx, ecx, times_4, 10000));
|
|
__ shl(edx, 1);
|
|
__ shl(edx, 6);
|
|
__ shl_cl(edx);
|
|
__ shl(Operand(ebx, ecx, times_4, 10000), 1);
|
|
__ shl(Operand(ebx, ecx, times_4, 10000), 6);
|
|
__ shl_cl(Operand(ebx, ecx, times_4, 10000));
|
|
__ shrd_cl(Operand(ebx, ecx, times_4, 10000), edx);
|
|
__ shr(edx, 1);
|
|
__ shr(edx, 7);
|
|
__ shr_cl(edx);
|
|
__ shr(Operand(ebx, ecx, times_4, 10000), 1);
|
|
__ shr(Operand(ebx, ecx, times_4, 10000), 6);
|
|
__ shr_cl(Operand(ebx, ecx, times_4, 10000));
|
|
|
|
|
|
// Immediates
|
|
|
|
__ adc(edx, 12345);
|
|
|
|
__ add(ebx, Immediate(12));
|
|
__ add(Operand(edx, ecx, times_4, 10000), Immediate(12));
|
|
|
|
__ and_(ebx, 12345);
|
|
|
|
__ cmp(ebx, 12345);
|
|
__ cmp(ebx, Immediate(12));
|
|
__ cmp(Operand(edx, ecx, times_4, 10000), Immediate(12));
|
|
__ cmpb(eax, Immediate(100));
|
|
|
|
__ or_(ebx, 12345);
|
|
|
|
__ sub(ebx, Immediate(12));
|
|
__ sub(Operand(edx, ecx, times_4, 10000), Immediate(12));
|
|
|
|
__ xor_(ebx, 12345);
|
|
|
|
__ imul(edx, ecx, 12);
|
|
__ imul(edx, ecx, 1000);
|
|
|
|
__ cld();
|
|
__ rep_movs();
|
|
__ rep_stos();
|
|
__ stos();
|
|
|
|
__ sub(edx, Operand(ebx, ecx, times_4, 10000));
|
|
__ sub(edx, ebx);
|
|
|
|
__ test(edx, Immediate(12345));
|
|
__ test(edx, Operand(ebx, ecx, times_8, 10000));
|
|
__ test(Operand(esi, edi, times_1, -20000000), Immediate(300000000));
|
|
__ test_b(edx, Operand(ecx, ebx, times_2, 1000));
|
|
__ test_b(Operand(eax, -20), Immediate(0x9A));
|
|
__ nop();
|
|
|
|
__ xor_(edx, 12345);
|
|
__ xor_(edx, Operand(ebx, ecx, times_8, 10000));
|
|
__ bts(Operand(ebx, ecx, times_8, 10000), edx);
|
|
__ hlt();
|
|
__ int3();
|
|
__ ret(0);
|
|
__ ret(8);
|
|
|
|
// Calls
|
|
|
|
Label L1, L2;
|
|
__ bind(&L1);
|
|
__ nop();
|
|
__ call(&L1);
|
|
__ call(&L2);
|
|
__ nop();
|
|
__ bind(&L2);
|
|
__ call(Operand(ebx, ecx, times_4, 10000));
|
|
__ nop();
|
|
Handle<Code> ic = BUILTIN_CODE(isolate, ArrayFrom);
|
|
__ call(ic, RelocInfo::CODE_TARGET);
|
|
__ nop();
|
|
__ call(FUNCTION_ADDR(DummyStaticFunction), RelocInfo::RUNTIME_ENTRY);
|
|
__ nop();
|
|
|
|
__ jmp(&L1);
|
|
__ jmp(Operand(ebx, ecx, times_4, 10000));
|
|
__ jmp(ic, RelocInfo::CODE_TARGET);
|
|
__ nop();
|
|
|
|
|
|
Label Ljcc;
|
|
__ nop();
|
|
// long jumps
|
|
__ j(overflow, &Ljcc);
|
|
__ j(no_overflow, &Ljcc);
|
|
__ j(below, &Ljcc);
|
|
__ j(above_equal, &Ljcc);
|
|
__ j(equal, &Ljcc);
|
|
__ j(not_equal, &Ljcc);
|
|
__ j(below_equal, &Ljcc);
|
|
__ j(above, &Ljcc);
|
|
__ j(sign, &Ljcc);
|
|
__ j(not_sign, &Ljcc);
|
|
__ j(parity_even, &Ljcc);
|
|
__ j(parity_odd, &Ljcc);
|
|
__ j(less, &Ljcc);
|
|
__ j(greater_equal, &Ljcc);
|
|
__ j(less_equal, &Ljcc);
|
|
__ j(greater, &Ljcc);
|
|
__ nop();
|
|
__ bind(&Ljcc);
|
|
// short jumps
|
|
__ j(overflow, &Ljcc);
|
|
__ j(no_overflow, &Ljcc);
|
|
__ j(below, &Ljcc);
|
|
__ j(above_equal, &Ljcc);
|
|
__ j(equal, &Ljcc);
|
|
__ j(not_equal, &Ljcc);
|
|
__ j(below_equal, &Ljcc);
|
|
__ j(above, &Ljcc);
|
|
__ j(sign, &Ljcc);
|
|
__ j(not_sign, &Ljcc);
|
|
__ j(parity_even, &Ljcc);
|
|
__ j(parity_odd, &Ljcc);
|
|
__ j(less, &Ljcc);
|
|
__ j(greater_equal, &Ljcc);
|
|
__ j(less_equal, &Ljcc);
|
|
__ j(greater, &Ljcc);
|
|
|
|
// 0xD9 instructions
|
|
__ nop();
|
|
|
|
__ fld(1);
|
|
__ fld1();
|
|
__ fldz();
|
|
__ fldpi();
|
|
__ fabs();
|
|
__ fchs();
|
|
__ fprem();
|
|
__ fprem1();
|
|
__ fincstp();
|
|
__ ftst();
|
|
__ fxch(3);
|
|
__ fld_s(Operand(ebx, ecx, times_4, 10000));
|
|
__ fstp_s(Operand(ebx, ecx, times_4, 10000));
|
|
__ ffree(3);
|
|
__ fld_d(Operand(ebx, ecx, times_4, 10000));
|
|
__ fstp_d(Operand(ebx, ecx, times_4, 10000));
|
|
__ nop();
|
|
|
|
__ fild_s(Operand(ebx, ecx, times_4, 10000));
|
|
__ fistp_s(Operand(ebx, ecx, times_4, 10000));
|
|
__ fild_d(Operand(ebx, ecx, times_4, 10000));
|
|
__ fistp_d(Operand(ebx, ecx, times_4, 10000));
|
|
__ fnstsw_ax();
|
|
__ nop();
|
|
__ fadd(3);
|
|
__ fsub(3);
|
|
__ fmul(3);
|
|
__ fdiv(3);
|
|
|
|
__ faddp(3);
|
|
__ fsubp(3);
|
|
__ fmulp(3);
|
|
__ fdivp(3);
|
|
__ fcompp();
|
|
__ fwait();
|
|
__ frndint();
|
|
__ fninit();
|
|
__ nop();
|
|
|
|
// SSE instruction
|
|
{
|
|
// Move operation
|
|
__ movaps(xmm0, xmm1);
|
|
__ movups(xmm0, xmm1);
|
|
__ movups(xmm0, Operand(edx, 4));
|
|
__ movups(Operand(edx, 4), xmm0);
|
|
__ shufps(xmm0, xmm0, 0x0);
|
|
__ cvtsd2ss(xmm0, xmm1);
|
|
__ cvtsd2ss(xmm0, Operand(ebx, ecx, times_4, 10000));
|
|
__ movq(xmm0, Operand(edx, 4));
|
|
|
|
// logic operation
|
|
__ andps(xmm0, xmm1);
|
|
__ andps(xmm0, Operand(ebx, ecx, times_4, 10000));
|
|
__ andnps(xmm0, xmm1);
|
|
__ andnps(xmm0, Operand(ebx, ecx, times_4, 10000));
|
|
__ orps(xmm0, xmm1);
|
|
__ orps(xmm0, Operand(ebx, ecx, times_4, 10000));
|
|
__ xorps(xmm0, xmm1);
|
|
__ xorps(xmm0, Operand(ebx, ecx, times_4, 10000));
|
|
|
|
// Arithmetic operation
|
|
__ addss(xmm1, xmm0);
|
|
__ addss(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ mulss(xmm1, xmm0);
|
|
__ mulss(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ subss(xmm1, xmm0);
|
|
__ subss(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ divss(xmm1, xmm0);
|
|
__ divss(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ maxss(xmm1, xmm0);
|
|
__ maxss(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ minss(xmm1, xmm0);
|
|
__ minss(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ sqrtss(xmm1, xmm0);
|
|
__ sqrtss(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ addps(xmm1, xmm0);
|
|
__ addps(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ subps(xmm1, xmm0);
|
|
__ subps(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ mulps(xmm1, xmm0);
|
|
__ mulps(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ divps(xmm1, xmm0);
|
|
__ divps(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ minps(xmm1, xmm0);
|
|
__ minps(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ maxps(xmm1, xmm0);
|
|
__ maxps(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ rcpps(xmm1, xmm0);
|
|
__ rcpps(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ sqrtps(xmm1, xmm0);
|
|
__ sqrtps(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ rsqrtps(xmm1, xmm0);
|
|
__ rsqrtps(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
|
|
__ cmpeqps(xmm5, xmm1);
|
|
__ cmpeqps(xmm5, Operand(ebx, ecx, times_4, 10000));
|
|
__ cmpltps(xmm5, xmm1);
|
|
__ cmpltps(xmm5, Operand(ebx, ecx, times_4, 10000));
|
|
__ cmpleps(xmm5, xmm1);
|
|
__ cmpleps(xmm5, Operand(ebx, ecx, times_4, 10000));
|
|
__ cmpunordps(xmm5, xmm1);
|
|
__ cmpunordps(xmm5, Operand(ebx, ecx, times_4, 10000));
|
|
__ cmpneqps(xmm5, xmm1);
|
|
__ cmpneqps(xmm5, Operand(ebx, ecx, times_4, 10000));
|
|
|
|
__ ucomiss(xmm0, xmm1);
|
|
__ ucomiss(xmm0, Operand(ebx, ecx, times_4, 10000));
|
|
}
|
|
{
|
|
__ cvttss2si(edx, Operand(ebx, ecx, times_4, 10000));
|
|
__ cvtsi2sd(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ cvtss2sd(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ cvtss2sd(xmm1, xmm0);
|
|
__ cvtdq2ps(xmm1, xmm0);
|
|
__ cvtdq2ps(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ cvttps2dq(xmm1, xmm0);
|
|
__ cvttps2dq(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ movsd(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ movsd(Operand(ebx, ecx, times_4, 10000), xmm1);
|
|
// 128 bit move instructions.
|
|
__ movdqa(xmm0, Operand(ebx, ecx, times_4, 10000));
|
|
__ movdqa(Operand(ebx, ecx, times_4, 10000), xmm0);
|
|
__ movdqu(xmm0, Operand(ebx, ecx, times_4, 10000));
|
|
__ movdqu(Operand(ebx, ecx, times_4, 10000), xmm0);
|
|
|
|
__ movapd(xmm0, xmm1);
|
|
__ movapd(xmm0, Operand(edx, 4));
|
|
|
|
__ movd(xmm0, edi);
|
|
__ movd(xmm0, Operand(ebx, ecx, times_4, 10000));
|
|
__ movd(eax, xmm1);
|
|
__ movd(Operand(ebx, ecx, times_4, 10000), xmm1);
|
|
|
|
__ addsd(xmm1, xmm0);
|
|
__ addsd(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ mulsd(xmm1, xmm0);
|
|
__ mulsd(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ subsd(xmm1, xmm0);
|
|
__ subsd(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ divsd(xmm1, xmm0);
|
|
__ divsd(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ minsd(xmm1, xmm0);
|
|
__ minsd(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ maxsd(xmm1, xmm0);
|
|
__ maxsd(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ sqrtsd(xmm1, xmm0);
|
|
__ sqrtsd(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ ucomisd(xmm0, xmm1);
|
|
__ cmpltsd(xmm0, xmm1);
|
|
|
|
__ andpd(xmm0, xmm1);
|
|
__ andpd(xmm0, Operand(ebx, ecx, times_4, 10000));
|
|
__ andnpd(xmm0, xmm1);
|
|
__ andnpd(xmm0, Operand(ebx, ecx, times_4, 10000));
|
|
__ orpd(xmm0, xmm1);
|
|
__ orpd(xmm0, Operand(ebx, ecx, times_4, 10000));
|
|
__ xorpd(xmm0, xmm1);
|
|
__ xorpd(xmm0, Operand(ebx, ecx, times_4, 10000));
|
|
__ addpd(xmm1, xmm0);
|
|
__ addpd(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ subpd(xmm1, xmm0);
|
|
__ subpd(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ mulpd(xmm1, xmm0);
|
|
__ mulpd(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ divpd(xmm1, xmm0);
|
|
__ divpd(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ minpd(xmm1, xmm0);
|
|
__ minpd(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ maxpd(xmm1, xmm0);
|
|
__ maxpd(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
|
|
__ cmpeqpd(xmm5, xmm1);
|
|
__ cmpeqpd(xmm5, Operand(ebx, ecx, times_4, 10000));
|
|
__ cmpltpd(xmm5, xmm1);
|
|
__ cmpltpd(xmm5, Operand(ebx, ecx, times_4, 10000));
|
|
__ cmplepd(xmm5, xmm1);
|
|
__ cmplepd(xmm5, Operand(ebx, ecx, times_4, 10000));
|
|
__ cmpunordpd(xmm5, xmm1);
|
|
__ cmpunordpd(xmm5, Operand(ebx, ecx, times_4, 10000));
|
|
__ cmpneqpd(xmm5, xmm1);
|
|
__ cmpneqpd(xmm5, Operand(ebx, ecx, times_4, 10000));
|
|
|
|
__ psllw(xmm0, 17);
|
|
__ pslld(xmm0, 17);
|
|
__ psrlw(xmm0, 17);
|
|
__ psrld(xmm0, 17);
|
|
__ psraw(xmm0, 17);
|
|
__ psrad(xmm0, 17);
|
|
__ psllq(xmm0, 17);
|
|
__ psrlq(xmm0, 17);
|
|
|
|
__ pshufhw(xmm5, xmm1, 5);
|
|
__ pshufhw(xmm5, Operand(edx, 4), 5);
|
|
__ pshuflw(xmm5, xmm1, 5);
|
|
__ pshuflw(xmm5, Operand(edx, 4), 5);
|
|
__ pshufd(xmm5, xmm1, 5);
|
|
__ pshufd(xmm5, Operand(edx, 4), 5);
|
|
__ pinsrw(xmm5, edx, 5);
|
|
__ pinsrw(xmm5, Operand(edx, 4), 5);
|
|
|
|
__ movmskps(edx, xmm5);
|
|
__ pmovmskb(edx, xmm5);
|
|
|
|
#define EMIT_SSE2_INSTR(instruction, notUsed1, notUsed2, notUsed3) \
|
|
__ instruction(xmm5, xmm1); \
|
|
__ instruction(xmm5, Operand(edx, 4));
|
|
|
|
SSE2_INSTRUCTION_LIST(EMIT_SSE2_INSTR)
|
|
#undef EMIT_SSE2_INSTR
|
|
}
|
|
|
|
// cmov.
|
|
{
|
|
__ cmov(overflow, eax, Operand(eax, 0));
|
|
__ cmov(no_overflow, eax, Operand(eax, 1));
|
|
__ cmov(below, eax, Operand(eax, 2));
|
|
__ cmov(above_equal, eax, Operand(eax, 3));
|
|
__ cmov(equal, eax, Operand(ebx, 0));
|
|
__ cmov(not_equal, eax, Operand(ebx, 1));
|
|
__ cmov(below_equal, eax, Operand(ebx, 2));
|
|
__ cmov(above, eax, Operand(ebx, 3));
|
|
__ cmov(sign, eax, Operand(ecx, 0));
|
|
__ cmov(not_sign, eax, Operand(ecx, 1));
|
|
__ cmov(parity_even, eax, Operand(ecx, 2));
|
|
__ cmov(parity_odd, eax, Operand(ecx, 3));
|
|
__ cmov(less, eax, Operand(edx, 0));
|
|
__ cmov(greater_equal, eax, Operand(edx, 1));
|
|
__ cmov(less_equal, eax, Operand(edx, 2));
|
|
__ cmov(greater, eax, Operand(edx, 3));
|
|
}
|
|
|
|
{
|
|
if (CpuFeatures::IsSupported(SSE3)) {
|
|
CpuFeatureScope scope(&assm, SSE3);
|
|
__ haddps(xmm1, xmm0);
|
|
__ haddps(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ movddup(xmm1, Operand(eax, 5));
|
|
__ movddup(xmm1, xmm2);
|
|
}
|
|
}
|
|
|
|
#define EMIT_SSE34_INSTR(instruction, notUsed1, notUsed2, notUsed3, notUsed4) \
|
|
__ instruction(xmm5, xmm1); \
|
|
__ instruction(xmm5, Operand(edx, 4));
|
|
|
|
{
|
|
if (CpuFeatures::IsSupported(SSSE3)) {
|
|
CpuFeatureScope scope(&assm, SSSE3);
|
|
SSSE3_INSTRUCTION_LIST(EMIT_SSE34_INSTR)
|
|
SSSE3_UNOP_INSTRUCTION_LIST(EMIT_SSE34_INSTR)
|
|
__ palignr(xmm5, xmm1, 5);
|
|
__ palignr(xmm5, Operand(edx, 4), 5);
|
|
}
|
|
}
|
|
|
|
{
|
|
if (CpuFeatures::IsSupported(SSE4_1)) {
|
|
CpuFeatureScope scope(&assm, SSE4_1);
|
|
__ pblendw(xmm5, xmm1, 5);
|
|
__ pblendw(xmm5, Operand(edx, 4), 5);
|
|
__ pextrb(eax, xmm0, 1);
|
|
__ pextrb(Operand(edx, 4), xmm0, 1);
|
|
__ pextrw(eax, xmm0, 1);
|
|
__ pextrw(Operand(edx, 4), xmm0, 1);
|
|
__ pextrd(eax, xmm0, 1);
|
|
__ pextrd(Operand(edx, 4), xmm0, 1);
|
|
__ insertps(xmm1, xmm2, 0);
|
|
__ insertps(xmm1, Operand(edx, 4), 0);
|
|
__ pinsrb(xmm1, eax, 0);
|
|
__ pinsrb(xmm1, Operand(edx, 4), 0);
|
|
__ pinsrd(xmm1, eax, 0);
|
|
__ pinsrd(xmm1, Operand(edx, 4), 0);
|
|
__ extractps(eax, xmm1, 0);
|
|
|
|
SSE4_INSTRUCTION_LIST(EMIT_SSE34_INSTR)
|
|
SSE4_RM_INSTRUCTION_LIST(EMIT_SSE34_INSTR)
|
|
}
|
|
}
|
|
#undef EMIT_SSE34_INSTR
|
|
|
|
// AVX instruction
|
|
{
|
|
if (CpuFeatures::IsSupported(AVX)) {
|
|
CpuFeatureScope scope(&assm, AVX);
|
|
__ vaddsd(xmm0, xmm1, xmm2);
|
|
__ vaddsd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vmulsd(xmm0, xmm1, xmm2);
|
|
__ vmulsd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vsubsd(xmm0, xmm1, xmm2);
|
|
__ vsubsd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vdivsd(xmm0, xmm1, xmm2);
|
|
__ vdivsd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vminsd(xmm0, xmm1, xmm2);
|
|
__ vminsd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vmaxsd(xmm0, xmm1, xmm2);
|
|
__ vmaxsd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vsqrtsd(xmm0, xmm1, xmm2);
|
|
__ vsqrtsd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
|
|
__ vaddss(xmm0, xmm1, xmm2);
|
|
__ vaddss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vmulss(xmm0, xmm1, xmm2);
|
|
__ vmulss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vsubss(xmm0, xmm1, xmm2);
|
|
__ vsubss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vdivss(xmm0, xmm1, xmm2);
|
|
__ vdivss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vminss(xmm0, xmm1, xmm2);
|
|
__ vminss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vmaxss(xmm0, xmm1, xmm2);
|
|
__ vmaxss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vsqrtss(xmm0, xmm1, xmm2);
|
|
__ vsqrtss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
|
|
__ vandps(xmm0, xmm1, xmm2);
|
|
__ vandps(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vandnps(xmm0, xmm1, xmm2);
|
|
__ vandnps(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vorps(xmm0, xmm1, xmm2);
|
|
__ vorps(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vxorps(xmm0, xmm1, xmm2);
|
|
__ vxorps(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vaddps(xmm0, xmm1, xmm2);
|
|
__ vaddps(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vmulps(xmm0, xmm1, xmm2);
|
|
__ vmulps(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vsubps(xmm0, xmm1, xmm2);
|
|
__ vsubps(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vminps(xmm0, xmm1, xmm2);
|
|
__ vminps(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vdivps(xmm0, xmm1, xmm2);
|
|
__ vdivps(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vmaxps(xmm0, xmm1, xmm2);
|
|
__ vmaxps(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vrcpps(xmm1, xmm0);
|
|
__ vrcpps(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vsqrtps(xmm1, xmm0);
|
|
__ vsqrtps(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vrsqrtps(xmm1, xmm0);
|
|
__ vrsqrtps(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vmovaps(xmm0, xmm1);
|
|
__ vmovapd(xmm0, xmm1);
|
|
__ vmovapd(xmm0, Operand(ebx, ecx, times_4, 10000));
|
|
__ vshufps(xmm0, xmm1, xmm2, 3);
|
|
__ vshufps(xmm0, xmm1, Operand(edx, 4), 3);
|
|
__ vhaddps(xmm0, xmm1, xmm2);
|
|
__ vhaddps(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
|
|
__ vcmpeqps(xmm5, xmm4, xmm1);
|
|
__ vcmpeqps(xmm5, xmm4, Operand(ebx, ecx, times_4, 10000));
|
|
__ vcmpltps(xmm5, xmm4, xmm1);
|
|
__ vcmpltps(xmm5, xmm4, Operand(ebx, ecx, times_4, 10000));
|
|
__ vcmpleps(xmm5, xmm4, xmm1);
|
|
__ vcmpleps(xmm5, xmm4, Operand(ebx, ecx, times_4, 10000));
|
|
__ vcmpunordps(xmm5, xmm4, xmm1);
|
|
__ vcmpunordps(xmm5, xmm4, Operand(ebx, ecx, times_4, 10000));
|
|
__ vcmpneqps(xmm5, xmm4, xmm1);
|
|
__ vcmpneqps(xmm5, xmm4, Operand(ebx, ecx, times_4, 10000));
|
|
|
|
__ vandpd(xmm0, xmm1, xmm2);
|
|
__ vandpd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vandnpd(xmm0, xmm1, xmm2);
|
|
__ vandnpd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vorpd(xmm0, xmm1, xmm2);
|
|
__ vorpd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vxorpd(xmm0, xmm1, xmm2);
|
|
__ vxorpd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vaddpd(xmm0, xmm1, xmm2);
|
|
__ vaddpd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vmulpd(xmm0, xmm1, xmm2);
|
|
__ vmulpd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vsubpd(xmm0, xmm1, xmm2);
|
|
__ vsubpd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vminpd(xmm0, xmm1, xmm2);
|
|
__ vminpd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vdivpd(xmm0, xmm1, xmm2);
|
|
__ vdivpd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vmaxpd(xmm0, xmm1, xmm2);
|
|
__ vmaxpd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
|
|
__ vcmpeqpd(xmm5, xmm4, xmm1);
|
|
__ vcmpeqpd(xmm5, xmm4, Operand(ebx, ecx, times_4, 10000));
|
|
__ vcmpltpd(xmm5, xmm4, xmm1);
|
|
__ vcmpltpd(xmm5, xmm4, Operand(ebx, ecx, times_4, 10000));
|
|
__ vcmplepd(xmm5, xmm4, xmm1);
|
|
__ vcmplepd(xmm5, xmm4, Operand(ebx, ecx, times_4, 10000));
|
|
__ vcmpunordpd(xmm5, xmm4, xmm1);
|
|
__ vcmpunordpd(xmm5, xmm4, Operand(ebx, ecx, times_4, 10000));
|
|
__ vcmpneqpd(xmm5, xmm4, xmm1);
|
|
__ vcmpneqpd(xmm5, xmm4, Operand(ebx, ecx, times_4, 10000));
|
|
|
|
__ vpsllw(xmm0, xmm7, 21);
|
|
__ vpslld(xmm0, xmm7, 21);
|
|
__ vpsllq(xmm0, xmm7, 21);
|
|
__ vpsrlw(xmm0, xmm7, 21);
|
|
__ vpsrld(xmm0, xmm7, 21);
|
|
__ vpsrlq(xmm0, xmm7, 21);
|
|
__ vpsraw(xmm0, xmm7, 21);
|
|
__ vpsrad(xmm0, xmm7, 21);
|
|
|
|
__ vpshufhw(xmm5, xmm1, 5);
|
|
__ vpshufhw(xmm5, Operand(edx, 4), 5);
|
|
__ vpshuflw(xmm5, xmm1, 5);
|
|
__ vpshuflw(xmm5, Operand(edx, 4), 5);
|
|
__ vpshufd(xmm5, xmm1, 5);
|
|
__ vpshufd(xmm5, Operand(edx, 4), 5);
|
|
__ vpblendw(xmm5, xmm1, xmm0, 5);
|
|
__ vpblendw(xmm5, xmm1, Operand(edx, 4), 5);
|
|
__ vpalignr(xmm5, xmm1, xmm0, 5);
|
|
__ vpalignr(xmm5, xmm1, Operand(edx, 4), 5);
|
|
__ vpextrb(eax, xmm0, 1);
|
|
__ vpextrb(Operand(edx, 4), xmm0, 1);
|
|
__ vpextrw(eax, xmm0, 1);
|
|
__ vpextrw(Operand(edx, 4), xmm0, 1);
|
|
__ vpextrd(eax, xmm0, 1);
|
|
__ vpextrd(Operand(edx, 4), xmm0, 1);
|
|
__ vinsertps(xmm0, xmm1, xmm2, 0);
|
|
__ vinsertps(xmm0, xmm1, Operand(edx, 4), 0);
|
|
__ vpinsrb(xmm0, xmm1, eax, 0);
|
|
__ vpinsrb(xmm0, xmm1, Operand(edx, 4), 0);
|
|
__ vpinsrw(xmm0, xmm1, eax, 0);
|
|
__ vpinsrw(xmm0, xmm1, Operand(edx, 4), 0);
|
|
__ vpinsrd(xmm0, xmm1, eax, 0);
|
|
__ vpinsrd(xmm0, xmm1, Operand(edx, 4), 0);
|
|
|
|
__ vcvtdq2ps(xmm1, xmm0);
|
|
__ vcvtdq2ps(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vcvttps2dq(xmm1, xmm0);
|
|
__ vcvttps2dq(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
|
|
__ vmovddup(xmm1, xmm2);
|
|
__ vmovddup(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vbroadcastss(xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vmovdqu(xmm0, Operand(ebx, ecx, times_4, 10000));
|
|
__ vmovdqu(Operand(ebx, ecx, times_4, 10000), xmm0);
|
|
__ vmovd(xmm0, edi);
|
|
__ vmovd(xmm0, Operand(ebx, ecx, times_4, 10000));
|
|
__ vmovd(eax, xmm1);
|
|
__ vmovd(Operand(ebx, ecx, times_4, 10000), xmm1);
|
|
|
|
__ vmovmskps(edx, xmm5);
|
|
__ vpmovmskb(ebx, xmm1);
|
|
|
|
#define EMIT_SSE2_AVXINSTR(instruction, notUsed1, notUsed2, notUsed3) \
|
|
__ v##instruction(xmm7, xmm5, xmm1); \
|
|
__ v##instruction(xmm7, xmm5, Operand(edx, 4));
|
|
|
|
SSE2_INSTRUCTION_LIST(EMIT_SSE2_AVXINSTR)
|
|
#undef EMIT_SSE2_AVXINSTR
|
|
|
|
#define EMIT_SSE34_AVXINSTR(instruction, notUsed1, notUsed2, notUsed3, \
|
|
notUsed4) \
|
|
__ v##instruction(xmm7, xmm5, xmm1); \
|
|
__ v##instruction(xmm7, xmm5, Operand(edx, 4));
|
|
|
|
SSSE3_INSTRUCTION_LIST(EMIT_SSE34_AVXINSTR)
|
|
SSE4_INSTRUCTION_LIST(EMIT_SSE34_AVXINSTR)
|
|
#undef EMIT_SSE34_AVXINSTR
|
|
|
|
#define EMIT_SSE4_RM_AVXINSTR(instruction, notUsed1, notUsed2, notUsed3, \
|
|
notUsed4) \
|
|
__ v##instruction(xmm5, xmm1); \
|
|
__ v##instruction(xmm5, Operand(edx, 4));
|
|
|
|
SSSE3_UNOP_INSTRUCTION_LIST(EMIT_SSE4_RM_AVXINSTR)
|
|
SSE4_RM_INSTRUCTION_LIST(EMIT_SSE4_RM_AVXINSTR)
|
|
#undef EMIT_SSE4_RM_AVXINSTR
|
|
}
|
|
}
|
|
|
|
// FMA3 instruction
|
|
{
|
|
if (CpuFeatures::IsSupported(FMA3)) {
|
|
CpuFeatureScope scope(&assm, FMA3);
|
|
__ vfmadd132sd(xmm0, xmm1, xmm2);
|
|
__ vfmadd132sd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vfmadd213sd(xmm0, xmm1, xmm2);
|
|
__ vfmadd213sd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vfmadd231sd(xmm0, xmm1, xmm2);
|
|
__ vfmadd231sd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
|
|
__ vfmsub132sd(xmm0, xmm1, xmm2);
|
|
__ vfmsub132sd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vfmsub213sd(xmm0, xmm1, xmm2);
|
|
__ vfmsub213sd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vfmsub231sd(xmm0, xmm1, xmm2);
|
|
__ vfmsub231sd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
|
|
__ vfnmadd132sd(xmm0, xmm1, xmm2);
|
|
__ vfnmadd132sd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vfnmadd213sd(xmm0, xmm1, xmm2);
|
|
__ vfnmadd213sd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vfnmadd231sd(xmm0, xmm1, xmm2);
|
|
__ vfnmadd231sd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
|
|
__ vfnmsub132sd(xmm0, xmm1, xmm2);
|
|
__ vfnmsub132sd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vfnmsub213sd(xmm0, xmm1, xmm2);
|
|
__ vfnmsub213sd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vfnmsub231sd(xmm0, xmm1, xmm2);
|
|
__ vfnmsub231sd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
|
|
__ vfmadd132ss(xmm0, xmm1, xmm2);
|
|
__ vfmadd132ss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vfmadd213ss(xmm0, xmm1, xmm2);
|
|
__ vfmadd213ss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vfmadd231ss(xmm0, xmm1, xmm2);
|
|
__ vfmadd231ss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
|
|
__ vfmsub132ss(xmm0, xmm1, xmm2);
|
|
__ vfmsub132ss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vfmsub213ss(xmm0, xmm1, xmm2);
|
|
__ vfmsub213ss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vfmsub231ss(xmm0, xmm1, xmm2);
|
|
__ vfmsub231ss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
|
|
__ vfnmadd132ss(xmm0, xmm1, xmm2);
|
|
__ vfnmadd132ss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vfnmadd213ss(xmm0, xmm1, xmm2);
|
|
__ vfnmadd213ss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vfnmadd231ss(xmm0, xmm1, xmm2);
|
|
__ vfnmadd231ss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
|
|
__ vfnmsub132ss(xmm0, xmm1, xmm2);
|
|
__ vfnmsub132ss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vfnmsub213ss(xmm0, xmm1, xmm2);
|
|
__ vfnmsub213ss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
__ vfnmsub231ss(xmm0, xmm1, xmm2);
|
|
__ vfnmsub231ss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000));
|
|
}
|
|
}
|
|
|
|
// BMI1 instructions
|
|
{
|
|
if (CpuFeatures::IsSupported(BMI1)) {
|
|
CpuFeatureScope scope(&assm, BMI1);
|
|
__ andn(eax, ebx, ecx);
|
|
__ andn(eax, ebx, Operand(ebx, ecx, times_4, 10000));
|
|
__ bextr(eax, ebx, ecx);
|
|
__ bextr(eax, Operand(ebx, ecx, times_4, 10000), ebx);
|
|
__ blsi(eax, ebx);
|
|
__ blsi(eax, Operand(ebx, ecx, times_4, 10000));
|
|
__ blsmsk(eax, ebx);
|
|
__ blsmsk(eax, Operand(ebx, ecx, times_4, 10000));
|
|
__ blsr(eax, ebx);
|
|
__ blsr(eax, Operand(ebx, ecx, times_4, 10000));
|
|
__ tzcnt(eax, ebx);
|
|
__ tzcnt(eax, Operand(ebx, ecx, times_4, 10000));
|
|
}
|
|
}
|
|
|
|
// LZCNT instructions
|
|
{
|
|
if (CpuFeatures::IsSupported(LZCNT)) {
|
|
CpuFeatureScope scope(&assm, LZCNT);
|
|
__ lzcnt(eax, ebx);
|
|
__ lzcnt(eax, Operand(ebx, ecx, times_4, 10000));
|
|
}
|
|
}
|
|
|
|
// POPCNT instructions
|
|
{
|
|
if (CpuFeatures::IsSupported(POPCNT)) {
|
|
CpuFeatureScope scope(&assm, POPCNT);
|
|
__ popcnt(eax, ebx);
|
|
__ popcnt(eax, Operand(ebx, ecx, times_4, 10000));
|
|
}
|
|
}
|
|
|
|
// BMI2 instructions
|
|
{
|
|
if (CpuFeatures::IsSupported(BMI2)) {
|
|
CpuFeatureScope scope(&assm, BMI2);
|
|
__ bzhi(eax, ebx, ecx);
|
|
__ bzhi(eax, Operand(ebx, ecx, times_4, 10000), ebx);
|
|
__ mulx(eax, ebx, ecx);
|
|
__ mulx(eax, ebx, Operand(ebx, ecx, times_4, 10000));
|
|
__ pdep(eax, ebx, ecx);
|
|
__ pdep(eax, ebx, Operand(ebx, ecx, times_4, 10000));
|
|
__ pext(eax, ebx, ecx);
|
|
__ pext(eax, ebx, Operand(ebx, ecx, times_4, 10000));
|
|
__ sarx(eax, ebx, ecx);
|
|
__ sarx(eax, Operand(ebx, ecx, times_4, 10000), ebx);
|
|
__ shlx(eax, ebx, ecx);
|
|
__ shlx(eax, Operand(ebx, ecx, times_4, 10000), ebx);
|
|
__ shrx(eax, ebx, ecx);
|
|
__ shrx(eax, Operand(ebx, ecx, times_4, 10000), ebx);
|
|
__ rorx(eax, ebx, 31);
|
|
__ rorx(eax, Operand(ebx, ecx, times_4, 10000), 31);
|
|
}
|
|
}
|
|
|
|
// xadd.
|
|
{
|
|
__ xadd(Operand(eax, 8), eax);
|
|
__ xadd_w(Operand(ebx, 8), eax);
|
|
__ xadd_b(Operand(ebx, 8), eax);
|
|
}
|
|
// xchg.
|
|
{
|
|
__ xchg_b(eax, Operand(eax, 8));
|
|
__ xchg_w(eax, Operand(ebx, 8));
|
|
__ xchg(eax, eax);
|
|
__ xchg(eax, ebx);
|
|
__ xchg(ebx, ebx);
|
|
__ xchg(ebx, Operand(esp, 12));
|
|
}
|
|
|
|
// cmpxchg.
|
|
{
|
|
__ cmpxchg_b(Operand(esp, 12), eax);
|
|
__ cmpxchg_w(Operand(ebx, ecx, times_4, 10000), eax);
|
|
__ cmpxchg(Operand(ebx, ecx, times_4, 10000), eax);
|
|
__ cmpxchg(Operand(ebx, ecx, times_4, 10000), eax);
|
|
__ cmpxchg8b(Operand(ebx, ecx, times_8, 10000));
|
|
}
|
|
|
|
// lock prefix.
|
|
{
|
|
__ lock();
|
|
__ cmpxchg(Operand(esp, 12), ebx);
|
|
|
|
__ lock();
|
|
__ xchg_w(eax, Operand(ecx, 8));
|
|
}
|
|
|
|
// Nop instructions
|
|
for (int i = 0; i < 16; i++) {
|
|
__ Nop(i);
|
|
}
|
|
|
|
__ mfence();
|
|
__ lfence();
|
|
__ pause();
|
|
__ ret(0);
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build();
|
|
USE(code);
|
|
#ifdef OBJECT_PRINT
|
|
StdoutStream os;
|
|
code->Print(os);
|
|
Address begin = code->raw_instruction_start();
|
|
Address end = code->raw_instruction_end();
|
|
disasm::Disassembler::Disassemble(stdout, reinterpret_cast<byte*>(begin),
|
|
reinterpret_cast<byte*>(end));
|
|
#endif
|
|
}
|
|
|
|
#undef __
|
|
|
|
} // namespace internal
|
|
} // namespace v8
|