506c09797c
In AVX, it is better to use the appropriate integer or floating point moves depending on which instructions produce/consume these moves, since there can be a delay moving from integer to floating point domain. On SSE systems, it is less important, and we can move movaps/movups which is 1 byte shorter than movdqa/movdqu. This patch cleans up a couple of places, and defines macro-assembler functions Movdqa, Movdqu, Movapd, to call into movaps/movups when AVX is not supported. Change-Id: Iba6c54e218875f1a70f61792978d7b3f69edfb4b Reviewed-on: https://chromium-review.googlesource.com/c/v8/v8/+/2599843 Commit-Queue: Zhi An Ng <zhin@chromium.org> Reviewed-by: Bill Budge <bbudge@chromium.org> Cr-Commit-Position: refs/heads/master@{#71884}
1039 lines
34 KiB
C++
1039 lines
34 KiB
C++
// Copyright 2011 the V8 project authors. All rights reserved.
|
|
// Redistribution and use in source and binary forms, with or without
|
|
// modification, are permitted provided that the following conditions are
|
|
// met:
|
|
//
|
|
// * Redistributions of source code must retain the above copyright
|
|
// notice, this list of conditions and the following disclaimer.
|
|
// * Redistributions in binary form must reproduce the above
|
|
// copyright notice, this list of conditions and the following
|
|
// disclaimer in the documentation and/or other materials provided
|
|
// with the distribution.
|
|
// * Neither the name of Google Inc. nor the names of its
|
|
// contributors may be used to endorse or promote products derived
|
|
// from this software without specific prior written permission.
|
|
//
|
|
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
#include <stdlib.h>
|
|
|
|
#include "src/init/v8.h"
|
|
|
|
#include "src/codegen/code-factory.h"
|
|
#include "src/codegen/macro-assembler.h"
|
|
#include "src/debug/debug.h"
|
|
#include "src/diagnostics/disasm.h"
|
|
#include "src/diagnostics/disassembler.h"
|
|
#include "src/execution/frames-inl.h"
|
|
#include "src/utils/ostreams.h"
|
|
#include "src/objects/objects-inl.h"
|
|
#include "test/cctest/cctest.h"
|
|
|
|
namespace v8 {
|
|
namespace internal {
|
|
|
|
#define __ assm.
|
|
|
|
TEST(DisasmX64) {
|
|
CcTest::InitializeVM();
|
|
Isolate* isolate = CcTest::i_isolate();
|
|
HandleScope scope(isolate);
|
|
v8::internal::byte buffer[8192];
|
|
Assembler assm(AssemblerOptions{},
|
|
ExternalAssemblerBuffer(buffer, sizeof buffer));
|
|
|
|
// Short immediate instructions
|
|
__ addq(rax, Immediate(12345678));
|
|
__ orq(rax, Immediate(12345678));
|
|
__ subq(rax, Immediate(12345678));
|
|
__ xorq(rax, Immediate(12345678));
|
|
__ andq(rax, Immediate(12345678));
|
|
|
|
// ---- This one caused crash
|
|
__ movq(rbx, Operand(rsp, rcx, times_2, 0)); // [rsp+rcx*4]
|
|
|
|
// ---- All instructions that I can think of
|
|
__ addq(rdx, rbx);
|
|
__ addq(rdx, Operand(rbx, 0));
|
|
__ addq(rdx, Operand(rbx, 16));
|
|
__ addq(rdx, Operand(rbx, 1999));
|
|
__ addq(rdx, Operand(rbx, -4));
|
|
__ addq(rdx, Operand(rbx, -1999));
|
|
__ addq(rdx, Operand(rsp, 0));
|
|
__ addq(rdx, Operand(rsp, 16));
|
|
__ addq(rdx, Operand(rsp, 1999));
|
|
__ addq(rdx, Operand(rsp, -4));
|
|
__ addq(rdx, Operand(rsp, -1999));
|
|
__ nop();
|
|
__ addq(rsi, Operand(rcx, times_4, 0));
|
|
__ addq(rsi, Operand(rcx, times_4, 24));
|
|
__ addq(rsi, Operand(rcx, times_4, -4));
|
|
__ addq(rsi, Operand(rcx, times_4, -1999));
|
|
__ nop();
|
|
__ addq(rdi, Operand(rbp, rcx, times_4, 0));
|
|
__ addq(rdi, Operand(rbp, rcx, times_4, 12));
|
|
__ addq(rdi, Operand(rbp, rcx, times_4, -8));
|
|
__ addq(rdi, Operand(rbp, rcx, times_4, -3999));
|
|
__ addq(Operand(rbp, rcx, times_4, 12), Immediate(12));
|
|
|
|
__ bswapl(rax);
|
|
__ bswapq(rdi);
|
|
__ bsrl(rax, r15);
|
|
__ bsrl(r9, Operand(rcx, times_8, 91919));
|
|
|
|
__ nop();
|
|
__ addq(rbx, Immediate(12));
|
|
__ nop();
|
|
__ nop();
|
|
__ andq(rdx, Immediate(3));
|
|
__ andq(rdx, Operand(rsp, 4));
|
|
__ cmpq(rdx, Immediate(3));
|
|
__ cmpq(rdx, Operand(rsp, 4));
|
|
__ cmpq(Operand(rbp, rcx, times_4, 0), Immediate(1000));
|
|
__ cmpb(rbx, Operand(rbp, rcx, times_2, 0));
|
|
__ cmpb(Operand(rbp, rcx, times_2, 0), rbx);
|
|
__ orq(rdx, Immediate(3));
|
|
__ xorq(rdx, Immediate(3));
|
|
__ nop();
|
|
__ cpuid();
|
|
__ movsxbl(rdx, Operand(rcx, 0));
|
|
__ movsxbq(rdx, Operand(rcx, 0));
|
|
__ movsxwl(rdx, Operand(rcx, 0));
|
|
__ movsxwq(rdx, Operand(rcx, 0));
|
|
__ movzxbl(rdx, Operand(rcx, 0));
|
|
__ movzxwl(rdx, Operand(rcx, 0));
|
|
__ movzxbq(rdx, Operand(rcx, 0));
|
|
__ movzxwq(rdx, Operand(rcx, 0));
|
|
|
|
__ nop();
|
|
__ imulq(rdx, rcx);
|
|
__ shld(rdx, rcx);
|
|
__ shrd(rdx, rcx);
|
|
__ shlq(Operand(rdi, rax, times_4, 100), Immediate(1));
|
|
__ shlq(Operand(rdi, rax, times_4, 100), Immediate(6));
|
|
__ shlq(Operand(r15, 0), Immediate(1));
|
|
__ shlq(Operand(r15, 0), Immediate(6));
|
|
__ shlq_cl(Operand(r15, 0));
|
|
__ shlq_cl(Operand(r15, 0));
|
|
__ shlq_cl(Operand(rdi, rax, times_4, 100));
|
|
__ shlq_cl(Operand(rdi, rax, times_4, 100));
|
|
__ shlq(rdx, Immediate(1));
|
|
__ shlq(rdx, Immediate(6));
|
|
__ shll(Operand(rdi, rax, times_4, 100), Immediate(1));
|
|
__ shll(Operand(rdi, rax, times_4, 100), Immediate(6));
|
|
__ shll(Operand(r15, 0), Immediate(1));
|
|
__ shll(Operand(r15, 0), Immediate(6));
|
|
__ shll_cl(Operand(r15, 0));
|
|
__ shll_cl(Operand(r15, 0));
|
|
__ shll_cl(Operand(rdi, rax, times_4, 100));
|
|
__ shll_cl(Operand(rdi, rax, times_4, 100));
|
|
__ shll(rdx, Immediate(1));
|
|
__ shll(rdx, Immediate(6));
|
|
__ btq(Operand(rdx, 0), rcx);
|
|
__ btsq(Operand(rdx, 0), rcx);
|
|
__ btsq(Operand(rbx, rcx, times_4, 0), rcx);
|
|
__ btsq(rcx, Immediate(13));
|
|
__ btrq(rcx, Immediate(13));
|
|
__ nop();
|
|
__ pushq(Immediate(12));
|
|
__ pushq(Immediate(23456));
|
|
__ pushq(rcx);
|
|
__ pushq(rsi);
|
|
__ pushq(Operand(rbp, StandardFrameConstants::kFunctionOffset));
|
|
__ pushq(Operand(rbx, rcx, times_4, 0));
|
|
__ pushq(Operand(rbx, rcx, times_4, 0));
|
|
__ pushq(Operand(rbx, rcx, times_4, 10000));
|
|
__ popq(rdx);
|
|
__ popq(rax);
|
|
__ popq(Operand(rbx, rcx, times_4, 0));
|
|
__ nop();
|
|
|
|
__ addq(rdx, Operand(rsp, 16));
|
|
__ addq(rdx, rcx);
|
|
__ movb(rdx, Operand(rcx, 0));
|
|
__ movb(rcx, Immediate(6));
|
|
__ movb(Operand(rsp, 16), rdx);
|
|
__ movw(Operand(rsp, 16), rdx);
|
|
__ nop();
|
|
__ movsxwq(rdx, Operand(rsp, 12));
|
|
__ movsxbq(rdx, Operand(rsp, 12));
|
|
__ movsxlq(rdx, Operand(rsp, 12));
|
|
__ movzxwq(rdx, Operand(rsp, 12));
|
|
__ movzxbq(rdx, Operand(rsp, 12));
|
|
__ nop();
|
|
__ movq(rdx, Immediate(1234567));
|
|
__ movq(rdx, Operand(rsp, 12));
|
|
__ movq(Operand(rbx, rcx, times_4, 10000), Immediate(12345));
|
|
__ movq(Operand(rbx, rcx, times_4, 10000), rdx);
|
|
__ nop();
|
|
__ decb(rdx);
|
|
__ decb(Operand(rax, 10));
|
|
__ decb(Operand(rbx, rcx, times_4, 10000));
|
|
__ decq(rdx);
|
|
__ cdq();
|
|
|
|
__ repstosl();
|
|
__ repstosq();
|
|
|
|
__ nop();
|
|
__ idivq(rdx);
|
|
__ mull(rdx);
|
|
__ mulq(rdx);
|
|
|
|
__ negb(rdx);
|
|
__ negb(r10);
|
|
__ negw(rdx);
|
|
__ negl(rdx);
|
|
__ negq(rdx);
|
|
__ negb(Operand(rsp, 12));
|
|
__ negw(Operand(rsp, 12));
|
|
__ negl(Operand(rsp, 12));
|
|
__ negb(Operand(rsp, 12));
|
|
|
|
__ notq(rdx);
|
|
__ testq(Operand(rbx, rcx, times_4, 10000), rdx);
|
|
|
|
__ imulq(rdx, rcx, Immediate(12));
|
|
__ imulq(rdx, rcx, Immediate(1000));
|
|
__ imulq(rdx, Operand(rbx, rcx, times_4, 10000));
|
|
__ imulq(rdx, Operand(rbx, rcx, times_4, 10000), Immediate(12));
|
|
__ imulq(rdx, Operand(rbx, rcx, times_4, 10000), Immediate(1000));
|
|
__ imull(r15, rcx, Immediate(12));
|
|
__ imull(r15, rcx, Immediate(1000));
|
|
__ imull(r15, Operand(rbx, rcx, times_4, 10000));
|
|
__ imull(r15, Operand(rbx, rcx, times_4, 10000), Immediate(12));
|
|
__ imull(r15, Operand(rbx, rcx, times_4, 10000), Immediate(1000));
|
|
|
|
__ incq(rdx);
|
|
__ incq(Operand(rbx, rcx, times_4, 10000));
|
|
__ pushq(Operand(rbx, rcx, times_4, 10000));
|
|
__ popq(Operand(rbx, rcx, times_4, 10000));
|
|
__ jmp(Operand(rbx, rcx, times_4, 10000));
|
|
|
|
__ leaq(rdx, Operand(rbx, rcx, times_4, 10000));
|
|
__ orq(rdx, Immediate(12345));
|
|
__ orq(rdx, Operand(rbx, rcx, times_4, 10000));
|
|
|
|
__ nop();
|
|
|
|
__ rclq(rdx, Immediate(1));
|
|
__ rclq(rdx, Immediate(7));
|
|
__ rcrq(rdx, Immediate(1));
|
|
__ rcrq(rdx, Immediate(7));
|
|
__ sarq(rdx, Immediate(1));
|
|
__ sarq(rdx, Immediate(6));
|
|
__ sarq_cl(rdx);
|
|
__ sbbq(rdx, rbx);
|
|
__ shld(rdx, rbx);
|
|
__ shlq(rdx, Immediate(1));
|
|
__ shlq(rdx, Immediate(6));
|
|
__ shlq_cl(rdx);
|
|
__ shrd(rdx, rbx);
|
|
__ shrq(rdx, Immediate(1));
|
|
__ shrq(rdx, Immediate(7));
|
|
__ shrq_cl(rdx);
|
|
|
|
|
|
// Immediates
|
|
|
|
__ addq(rbx, Immediate(12));
|
|
__ addq(Operand(rdx, rcx, times_4, 10000), Immediate(12));
|
|
|
|
__ andq(rbx, Immediate(12345));
|
|
|
|
__ cmpq(rbx, Immediate(12345));
|
|
__ cmpq(rbx, Immediate(12));
|
|
__ cmpq(Operand(rdx, rcx, times_4, 10000), Immediate(12));
|
|
__ cmpb(rax, Immediate(100));
|
|
|
|
__ orq(rbx, Immediate(12345));
|
|
|
|
__ subq(rbx, Immediate(12));
|
|
__ subq(Operand(rdx, rcx, times_4, 10000), Immediate(12));
|
|
|
|
__ xorq(rbx, Immediate(12345));
|
|
|
|
__ imulq(rdx, rcx, Immediate(12));
|
|
__ imulq(rdx, rcx, Immediate(1000));
|
|
|
|
__ cld();
|
|
|
|
__ subq(rdx, Operand(rbx, rcx, times_4, 10000));
|
|
__ subq(rdx, rbx);
|
|
|
|
__ testq(rdx, Immediate(12345));
|
|
__ testq(Operand(rbx, rcx, times_8, 10000), rdx);
|
|
__ testb(Operand(rcx, rbx, times_2, 1000), rdx);
|
|
__ testb(Operand(rax, -20), Immediate(0x9A));
|
|
__ nop();
|
|
|
|
__ xorq(rdx, Immediate(12345));
|
|
__ xorq(rdx, Operand(rbx, rcx, times_8, 10000));
|
|
__ hlt();
|
|
__ int3();
|
|
__ ret(0);
|
|
__ ret(8);
|
|
|
|
// Calls
|
|
|
|
Label L1, L2;
|
|
__ bind(&L1);
|
|
__ nop();
|
|
__ call(&L1);
|
|
__ call(&L2);
|
|
__ nop();
|
|
__ bind(&L2);
|
|
__ call(rcx);
|
|
__ nop();
|
|
Handle<Code> ic = BUILTIN_CODE(isolate, ArrayFrom);
|
|
__ call(ic, RelocInfo::CODE_TARGET);
|
|
__ nop();
|
|
|
|
__ jmp(&L1);
|
|
__ jmp(Operand(rbx, rcx, times_4, 10000));
|
|
__ jmp(ic, RelocInfo::CODE_TARGET);
|
|
__ nop();
|
|
|
|
|
|
Label Ljcc;
|
|
__ nop();
|
|
// long jumps
|
|
__ j(overflow, &Ljcc);
|
|
__ j(no_overflow, &Ljcc);
|
|
__ j(below, &Ljcc);
|
|
__ j(above_equal, &Ljcc);
|
|
__ j(equal, &Ljcc);
|
|
__ j(not_equal, &Ljcc);
|
|
__ j(below_equal, &Ljcc);
|
|
__ j(above, &Ljcc);
|
|
__ j(sign, &Ljcc);
|
|
__ j(not_sign, &Ljcc);
|
|
__ j(parity_even, &Ljcc);
|
|
__ j(parity_odd, &Ljcc);
|
|
__ j(less, &Ljcc);
|
|
__ j(greater_equal, &Ljcc);
|
|
__ j(less_equal, &Ljcc);
|
|
__ j(greater, &Ljcc);
|
|
__ nop();
|
|
__ bind(&Ljcc);
|
|
// short jumps
|
|
__ j(overflow, &Ljcc);
|
|
__ j(no_overflow, &Ljcc);
|
|
__ j(below, &Ljcc);
|
|
__ j(above_equal, &Ljcc);
|
|
__ j(equal, &Ljcc);
|
|
__ j(not_equal, &Ljcc);
|
|
__ j(below_equal, &Ljcc);
|
|
__ j(above, &Ljcc);
|
|
__ j(sign, &Ljcc);
|
|
__ j(not_sign, &Ljcc);
|
|
__ j(parity_even, &Ljcc);
|
|
__ j(parity_odd, &Ljcc);
|
|
__ j(less, &Ljcc);
|
|
__ j(greater_equal, &Ljcc);
|
|
__ j(less_equal, &Ljcc);
|
|
__ j(greater, &Ljcc);
|
|
|
|
// 0xD9 instructions
|
|
__ nop();
|
|
|
|
__ fld(1);
|
|
__ fld1();
|
|
__ fldz();
|
|
__ fldpi();
|
|
__ fabs();
|
|
__ fchs();
|
|
__ fprem();
|
|
__ fprem1();
|
|
__ fincstp();
|
|
__ ftst();
|
|
__ fxch(3);
|
|
__ fld_s(Operand(rbx, rcx, times_4, 10000));
|
|
__ fstp_s(Operand(rbx, rcx, times_4, 10000));
|
|
__ ffree(3);
|
|
__ fld_d(Operand(rbx, rcx, times_4, 10000));
|
|
__ fstp_d(Operand(rbx, rcx, times_4, 10000));
|
|
__ nop();
|
|
|
|
__ fild_s(Operand(rbx, rcx, times_4, 10000));
|
|
__ fistp_s(Operand(rbx, rcx, times_4, 10000));
|
|
__ fild_d(Operand(rbx, rcx, times_4, 10000));
|
|
__ fistp_d(Operand(rbx, rcx, times_4, 10000));
|
|
__ fnstsw_ax();
|
|
__ nop();
|
|
__ fadd(3);
|
|
__ fsub(3);
|
|
__ fmul(3);
|
|
__ fdiv(3);
|
|
|
|
__ faddp(3);
|
|
__ fsubp(3);
|
|
__ fmulp(3);
|
|
__ fdivp(3);
|
|
__ fcompp();
|
|
__ fwait();
|
|
__ frndint();
|
|
__ fninit();
|
|
__ nop();
|
|
|
|
// SSE instruction
|
|
{
|
|
// Move operation
|
|
__ cvttss2si(rdx, Operand(rbx, rcx, times_4, 10000));
|
|
__ cvttss2si(rdx, xmm1);
|
|
__ cvtqsi2ss(xmm1, Operand(rbx, rcx, times_4, 10000));
|
|
__ cvtqsi2ss(xmm1, rdx);
|
|
__ cvttps2dq(xmm0, xmm1);
|
|
__ cvttps2dq(xmm0, Operand(rbx, rcx, times_4, 10000));
|
|
__ movaps(xmm0, xmm1);
|
|
__ movdqa(xmm0, Operand(rsp, 12));
|
|
__ movdqa(Operand(rsp, 12), xmm0);
|
|
__ movdqu(xmm0, Operand(rsp, 12));
|
|
__ movdqu(Operand(rsp, 12), xmm0);
|
|
__ movdqu(xmm1, xmm0);
|
|
__ movhlps(xmm5, xmm1);
|
|
__ movlps(xmm8, Operand(rbx, rcx, times_4, 10000));
|
|
__ movlps(Operand(rbx, rcx, times_4, 10000), xmm9);
|
|
__ movlhps(xmm5, xmm1);
|
|
__ movhps(xmm8, Operand(rbx, rcx, times_4, 10000));
|
|
__ movhps(Operand(rbx, rcx, times_4, 10000), xmm9);
|
|
__ shufps(xmm0, xmm9, 0x0);
|
|
|
|
__ ucomiss(xmm0, xmm1);
|
|
__ ucomiss(xmm0, Operand(rbx, rcx, times_4, 10000));
|
|
|
|
__ movmskps(rdx, xmm9);
|
|
|
|
#define EMIT_SSE_INSTR(instruction, notUsed1, notUsed2) \
|
|
__ instruction(xmm1, xmm0); \
|
|
__ instruction(xmm1, Operand(rbx, rcx, times_4, 10000));
|
|
SSE_BINOP_INSTRUCTION_LIST(EMIT_SSE_INSTR)
|
|
SSE_UNOP_INSTRUCTION_LIST(EMIT_SSE_INSTR)
|
|
#undef EMIT_SSE_INSTR
|
|
|
|
#define EMIT_SSE_INSTR(instruction, notUsed1, notUsed2, notUse3) \
|
|
__ instruction(xmm1, xmm0); \
|
|
__ instruction(xmm1, Operand(rbx, rcx, times_4, 10000));
|
|
SSE_INSTRUCTION_LIST_SS(EMIT_SSE_INSTR)
|
|
#undef EMIT_SSE_INSTR
|
|
}
|
|
|
|
// SSE2 instructions
|
|
{
|
|
__ cvttsd2si(rdx, Operand(rbx, rcx, times_4, 10000));
|
|
__ cvttsd2si(rdx, xmm1);
|
|
__ cvttsd2siq(rdx, xmm1);
|
|
__ cvttsd2siq(rdx, Operand(rbx, rcx, times_4, 10000));
|
|
__ cvtlsi2sd(xmm1, Operand(rbx, rcx, times_4, 10000));
|
|
__ cvtlsi2sd(xmm1, rdx);
|
|
__ cvtqsi2sd(xmm1, Operand(rbx, rcx, times_4, 10000));
|
|
__ cvtqsi2sd(xmm1, rdx);
|
|
__ cvtss2sd(xmm1, xmm9);
|
|
__ cvtss2sd(xmm1, Operand(rbx, rcx, times_4, 10000));
|
|
__ cvtsd2si(rdx, xmm9);
|
|
__ cvtsd2siq(rdx, xmm9);
|
|
|
|
__ movsd(xmm1, Operand(rbx, rcx, times_4, 10000));
|
|
__ movsd(Operand(rbx, rcx, times_4, 10000), xmm1);
|
|
// 128 bit move instructions.
|
|
__ movupd(xmm0, Operand(rbx, rcx, times_4, 10000));
|
|
__ movupd(Operand(rbx, rcx, times_4, 10000), xmm0);
|
|
__ movdqa(xmm0, Operand(rbx, rcx, times_4, 10000));
|
|
__ movdqa(Operand(rbx, rcx, times_4, 10000), xmm0);
|
|
__ movdqa(xmm0, xmm1);
|
|
|
|
__ ucomisd(xmm0, xmm1);
|
|
__ ucomisd(xmm8, Operand(rbx, rdx, times_4, 10000));
|
|
|
|
__ cmpltsd(xmm3, xmm11);
|
|
|
|
__ movmskpd(rdx, xmm9);
|
|
__ pmovmskb(rdx, xmm9);
|
|
|
|
__ pcmpeqd(xmm1, xmm0);
|
|
|
|
__ punpckldq(xmm1, xmm11);
|
|
__ punpckldq(xmm5, Operand(rdx, 4));
|
|
__ punpckhdq(xmm8, xmm15);
|
|
|
|
__ pshuflw(xmm2, xmm4, 3);
|
|
__ pshufhw(xmm1, xmm9, 6);
|
|
|
|
#define EMIT_SSE2_INSTR(instruction, notUsed1, notUsed2, notUsed3) \
|
|
__ instruction(xmm5, xmm1); \
|
|
__ instruction(xmm5, Operand(rdx, 4));
|
|
|
|
SSE2_INSTRUCTION_LIST(EMIT_SSE2_INSTR)
|
|
SSE2_UNOP_INSTRUCTION_LIST(EMIT_SSE2_INSTR)
|
|
SSE2_INSTRUCTION_LIST_SD(EMIT_SSE2_INSTR)
|
|
#undef EMIT_SSE2_INSTR
|
|
|
|
#define EMIT_SSE2_SHIFT_IMM(instruction, notUsed1, notUsed2, notUsed3, \
|
|
notUsed4) \
|
|
__ instruction(xmm3, 0xA3);
|
|
SSE2_INSTRUCTION_LIST_SHIFT_IMM(EMIT_SSE2_SHIFT_IMM)
|
|
#undef EMIT_SSE2_SHIFT_IMM
|
|
}
|
|
|
|
// cmov.
|
|
{
|
|
__ cmovq(overflow, rax, Operand(rax, 0));
|
|
__ cmovq(no_overflow, rax, Operand(rax, 1));
|
|
__ cmovq(below, rax, Operand(rax, 2));
|
|
__ cmovq(above_equal, rax, Operand(rax, 3));
|
|
__ cmovq(equal, rax, Operand(rbx, 0));
|
|
__ cmovq(not_equal, rax, Operand(rbx, 1));
|
|
__ cmovq(below_equal, rax, Operand(rbx, 2));
|
|
__ cmovq(above, rax, Operand(rbx, 3));
|
|
__ cmovq(sign, rax, Operand(rcx, 0));
|
|
__ cmovq(not_sign, rax, Operand(rcx, 1));
|
|
__ cmovq(parity_even, rax, Operand(rcx, 2));
|
|
__ cmovq(parity_odd, rax, Operand(rcx, 3));
|
|
__ cmovq(less, rax, Operand(rdx, 0));
|
|
__ cmovq(greater_equal, rax, Operand(rdx, 1));
|
|
__ cmovq(less_equal, rax, Operand(rdx, 2));
|
|
__ cmovq(greater, rax, Operand(rdx, 3));
|
|
}
|
|
|
|
{
|
|
if (CpuFeatures::IsSupported(SSE3)) {
|
|
CpuFeatureScope scope(&assm, SSE3);
|
|
__ haddps(xmm1, xmm0);
|
|
__ haddps(xmm1, Operand(rbx, rcx, times_4, 10000));
|
|
__ lddqu(xmm1, Operand(rdx, 4));
|
|
__ movddup(xmm1, Operand(rax, 5));
|
|
__ movddup(xmm1, xmm2);
|
|
__ movshdup(xmm1, xmm2);
|
|
}
|
|
}
|
|
|
|
#define EMIT_SSE34_INSTR(instruction, notUsed1, notUsed2, notUsed3, notUsed4) \
|
|
__ instruction(xmm5, xmm1); \
|
|
__ instruction(xmm5, Operand(rdx, 4));
|
|
|
|
#define EMIT_SSE34_IMM_INSTR(instruction, notUsed1, notUsed2, notUsed3, \
|
|
notUsed4) \
|
|
__ instruction(rbx, xmm15, 0); \
|
|
__ instruction(Operand(rax, 10), xmm0, 1);
|
|
|
|
{
|
|
if (CpuFeatures::IsSupported(SSSE3)) {
|
|
CpuFeatureScope scope(&assm, SSSE3);
|
|
__ palignr(xmm5, xmm1, 5);
|
|
__ palignr(xmm5, Operand(rdx, 4), 5);
|
|
SSSE3_INSTRUCTION_LIST(EMIT_SSE34_INSTR)
|
|
SSSE3_UNOP_INSTRUCTION_LIST(EMIT_SSE34_INSTR)
|
|
}
|
|
}
|
|
|
|
{
|
|
if (CpuFeatures::IsSupported(SSE4_1)) {
|
|
CpuFeatureScope scope(&assm, SSE4_1);
|
|
__ insertps(xmm5, xmm1, 123);
|
|
__ pinsrw(xmm2, rcx, 1);
|
|
__ pextrq(r12, xmm0, 1);
|
|
__ pinsrd(xmm9, r9, 0);
|
|
__ pinsrd(xmm5, Operand(rax, 4), 1);
|
|
__ pinsrq(xmm9, r9, 0);
|
|
__ pinsrq(xmm5, Operand(rax, 4), 1);
|
|
__ pblendw(xmm5, xmm1, 1);
|
|
__ pblendw(xmm9, Operand(rax, 4), 1);
|
|
|
|
__ cmpps(xmm5, xmm1, 1);
|
|
__ cmpps(xmm5, Operand(rbx, rcx, times_4, 10000), 1);
|
|
__ cmpeqps(xmm5, xmm1);
|
|
__ cmpeqps(xmm5, Operand(rbx, rcx, times_4, 10000));
|
|
__ cmpltps(xmm5, xmm1);
|
|
__ cmpltps(xmm5, Operand(rbx, rcx, times_4, 10000));
|
|
__ cmpleps(xmm5, xmm1);
|
|
__ cmpleps(xmm5, Operand(rbx, rcx, times_4, 10000));
|
|
__ cmpneqps(xmm5, xmm1);
|
|
__ cmpneqps(xmm5, Operand(rbx, rcx, times_4, 10000));
|
|
__ cmpnltps(xmm5, xmm1);
|
|
__ cmpnltps(xmm5, Operand(rbx, rcx, times_4, 10000));
|
|
__ cmpnleps(xmm5, xmm1);
|
|
__ cmpnleps(xmm5, Operand(rbx, rcx, times_4, 10000));
|
|
__ cmppd(xmm5, xmm1, 1);
|
|
__ cmppd(xmm5, Operand(rbx, rcx, times_4, 10000), 1);
|
|
__ cmpeqpd(xmm5, xmm1);
|
|
__ cmpeqpd(xmm5, Operand(rbx, rcx, times_4, 10000));
|
|
__ cmpltpd(xmm5, xmm1);
|
|
__ cmpltpd(xmm5, Operand(rbx, rcx, times_4, 10000));
|
|
__ cmplepd(xmm5, xmm1);
|
|
__ cmplepd(xmm5, Operand(rbx, rcx, times_4, 10000));
|
|
__ cmpneqpd(xmm5, xmm1);
|
|
__ cmpneqpd(xmm5, Operand(rbx, rcx, times_4, 10000));
|
|
__ cmpnltpd(xmm5, xmm1);
|
|
__ cmpnltpd(xmm5, Operand(rbx, rcx, times_4, 10000));
|
|
__ cmpnlepd(xmm5, xmm1);
|
|
__ cmpnlepd(xmm5, Operand(rbx, rcx, times_4, 10000));
|
|
|
|
__ movups(xmm5, xmm1);
|
|
__ movups(xmm5, Operand(rdx, 4));
|
|
__ movups(Operand(rdx, 4), xmm5);
|
|
__ pmulld(xmm5, xmm1);
|
|
__ pmulld(xmm5, Operand(rdx, 4));
|
|
__ pmullw(xmm5, xmm1);
|
|
__ pmullw(xmm5, Operand(rdx, 4));
|
|
__ pmuludq(xmm5, xmm1);
|
|
__ pmuludq(xmm5, Operand(rdx, 4));
|
|
__ psrldq(xmm5, 123);
|
|
__ pshufd(xmm5, xmm1, 3);
|
|
__ cvtps2dq(xmm5, xmm1);
|
|
__ cvtps2dq(xmm5, Operand(rdx, 4));
|
|
__ cvtdq2ps(xmm5, xmm1);
|
|
__ cvtdq2ps(xmm5, Operand(rdx, 4));
|
|
|
|
__ pblendvb(xmm5, xmm1);
|
|
__ blendvps(xmm5, xmm1);
|
|
__ blendvps(xmm5, Operand(rdx, 4));
|
|
__ blendvpd(xmm5, xmm1);
|
|
__ blendvpd(xmm5, Operand(rdx, 4));
|
|
|
|
__ roundps(xmm8, xmm3, kRoundUp);
|
|
__ roundpd(xmm8, xmm3, kRoundToNearest);
|
|
__ roundss(xmm8, xmm3, kRoundDown);
|
|
__ roundsd(xmm8, xmm3, kRoundDown);
|
|
|
|
SSE4_INSTRUCTION_LIST(EMIT_SSE34_INSTR)
|
|
SSE4_UNOP_INSTRUCTION_LIST(EMIT_SSE34_INSTR)
|
|
SSE4_EXTRACT_INSTRUCTION_LIST(EMIT_SSE34_IMM_INSTR)
|
|
}
|
|
}
|
|
|
|
{
|
|
if (CpuFeatures::IsSupported(SSE4_2)) {
|
|
CpuFeatureScope scope(&assm, SSE4_2);
|
|
|
|
SSE4_2_INSTRUCTION_LIST(EMIT_SSE34_INSTR)
|
|
}
|
|
}
|
|
#undef EMIT_SSE34_INSTR
|
|
#undef EMIT_SSE34_IMM_INSTR
|
|
|
|
// AVX instruction
|
|
{
|
|
if (CpuFeatures::IsSupported(AVX)) {
|
|
CpuFeatureScope scope(&assm, AVX);
|
|
__ vmovss(xmm6, xmm14, xmm2);
|
|
__ vmovss(xmm9, Operand(rbx, rcx, times_4, 10000));
|
|
__ vmovss(Operand(rbx, rcx, times_4, 10000), xmm0);
|
|
|
|
__ vaddss(xmm0, xmm1, xmm2);
|
|
__ vaddss(xmm0, xmm1, Operand(rbx, rcx, times_4, 10000));
|
|
__ vmulss(xmm0, xmm1, xmm2);
|
|
__ vmulss(xmm0, xmm1, Operand(rbx, rcx, times_4, 10000));
|
|
__ vsubss(xmm0, xmm1, xmm2);
|
|
__ vsubss(xmm0, xmm1, Operand(rbx, rcx, times_4, 10000));
|
|
__ vdivss(xmm0, xmm1, xmm2);
|
|
__ vdivss(xmm0, xmm1, Operand(rbx, rcx, times_2, 10000));
|
|
__ vminss(xmm8, xmm1, xmm2);
|
|
__ vminss(xmm9, xmm1, Operand(rbx, rcx, times_8, 10000));
|
|
__ vmaxss(xmm8, xmm1, xmm2);
|
|
__ vmaxss(xmm9, xmm1, Operand(rbx, rcx, times_1, 10000));
|
|
__ vsqrtss(xmm8, xmm1, xmm2);
|
|
__ vsqrtss(xmm9, xmm1, Operand(rbx, rcx, times_1, 10000));
|
|
__ vmovss(xmm9, Operand(r11, rcx, times_8, -10000));
|
|
__ vmovss(Operand(rbx, r9, times_4, 10000), xmm1);
|
|
__ vucomiss(xmm9, xmm1);
|
|
__ vucomiss(xmm8, Operand(rbx, rdx, times_2, 10981));
|
|
|
|
__ vmovd(xmm5, rdi);
|
|
__ vmovd(xmm9, Operand(rbx, rcx, times_4, 10000));
|
|
__ vmovd(r9, xmm6);
|
|
__ vmovq(xmm5, rdi);
|
|
__ vmovq(xmm9, Operand(rbx, rcx, times_4, 10000));
|
|
__ vmovq(r9, xmm6);
|
|
|
|
__ vmovsd(xmm6, xmm14, xmm2);
|
|
__ vmovsd(xmm9, Operand(rbx, rcx, times_4, 10000));
|
|
__ vmovsd(Operand(rbx, rcx, times_4, 10000), xmm0);
|
|
|
|
__ vmovdqa(xmm4, xmm5);
|
|
|
|
__ vmovdqu(xmm9, Operand(rbx, rcx, times_4, 10000));
|
|
__ vmovdqu(Operand(rbx, rcx, times_4, 10000), xmm0);
|
|
__ vmovdqu(xmm4, xmm5);
|
|
|
|
__ vmovhlps(xmm1, xmm3, xmm5);
|
|
__ vmovlps(xmm8, xmm9, Operand(rbx, rcx, times_4, 10000));
|
|
__ vmovlps(Operand(rbx, rcx, times_4, 10000), xmm9);
|
|
__ vmovlhps(xmm1, xmm3, xmm5);
|
|
__ vmovhps(xmm8, xmm9, Operand(rbx, rcx, times_4, 10000));
|
|
__ vmovhps(Operand(rbx, rcx, times_4, 10000), xmm12);
|
|
|
|
__ vroundps(xmm9, xmm2, kRoundUp);
|
|
__ vroundpd(xmm9, xmm2, kRoundToNearest);
|
|
__ vroundss(xmm9, xmm1, xmm2, kRoundDown);
|
|
__ vroundsd(xmm8, xmm3, xmm0, kRoundDown);
|
|
__ vucomisd(xmm9, xmm1);
|
|
__ vucomisd(xmm8, Operand(rbx, rdx, times_2, 10981));
|
|
|
|
__ vcvtss2sd(xmm4, xmm9, xmm11);
|
|
__ vcvtss2sd(xmm4, xmm9, Operand(rbx, rcx, times_1, 10000));
|
|
__ vcvttps2dq(xmm4, xmm11);
|
|
__ vcvtlsi2sd(xmm5, xmm9, rcx);
|
|
__ vcvtlsi2sd(xmm9, xmm3, Operand(rbx, r9, times_4, 10000));
|
|
__ vcvtqsi2sd(xmm5, xmm9, r11);
|
|
__ vcvttsd2si(r9, xmm6);
|
|
__ vcvttsd2si(rax, Operand(rbx, r9, times_4, 10000));
|
|
__ vcvttsd2siq(rdi, xmm9);
|
|
__ vcvttsd2siq(r8, Operand(r9, rbx, times_4, 10000));
|
|
__ vcvtsd2si(rdi, xmm9);
|
|
|
|
__ vmovaps(xmm10, xmm11);
|
|
__ vmovapd(xmm7, xmm0);
|
|
__ vmovupd(xmm0, Operand(rbx, rcx, times_4, 10000));
|
|
__ vmovupd(Operand(rbx, rcx, times_4, 10000), xmm0);
|
|
__ vmovmskpd(r9, xmm4);
|
|
__ vpmovmskb(r10, xmm9);
|
|
|
|
__ vmovups(xmm5, xmm1);
|
|
__ vmovups(xmm5, Operand(rdx, 4));
|
|
__ vmovups(Operand(rdx, 4), xmm5);
|
|
|
|
__ vandps(xmm0, xmm9, xmm2);
|
|
__ vandps(xmm9, xmm1, Operand(rbx, rcx, times_4, 10000));
|
|
__ vandnps(xmm0, xmm9, xmm2);
|
|
__ vandnps(xmm9, xmm1, Operand(rbx, rcx, times_4, 10000));
|
|
__ vxorps(xmm0, xmm1, xmm9);
|
|
__ vxorps(xmm0, xmm1, Operand(rbx, rcx, times_4, 10000));
|
|
__ vhaddps(xmm0, xmm1, xmm9);
|
|
__ vhaddps(xmm0, xmm1, Operand(rbx, rcx, times_4, 10000));
|
|
|
|
__ vpcmpeqd(xmm0, xmm15, xmm5);
|
|
__ vpcmpeqd(xmm15, xmm0, Operand(rbx, rcx, times_4, 10000));
|
|
|
|
__ vcmpps(xmm5, xmm4, xmm1, 1);
|
|
__ vcmpps(xmm5, xmm4, Operand(rbx, rcx, times_4, 10000), 1);
|
|
__ vcmpeqps(xmm5, xmm4, xmm1);
|
|
__ vcmpeqps(xmm5, xmm4, Operand(rbx, rcx, times_4, 10000));
|
|
__ vcmpltps(xmm5, xmm4, xmm1);
|
|
__ vcmpltps(xmm5, xmm4, Operand(rbx, rcx, times_4, 10000));
|
|
__ vcmpleps(xmm5, xmm4, xmm1);
|
|
__ vcmpleps(xmm5, xmm4, Operand(rbx, rcx, times_4, 10000));
|
|
__ vcmpneqps(xmm5, xmm4, xmm1);
|
|
__ vcmpneqps(xmm5, xmm4, Operand(rbx, rcx, times_4, 10000));
|
|
__ vcmpnltps(xmm5, xmm4, xmm1);
|
|
__ vcmpnltps(xmm5, xmm4, Operand(rbx, rcx, times_4, 10000));
|
|
__ vcmpnleps(xmm5, xmm4, xmm1);
|
|
__ vcmpnleps(xmm5, xmm4, Operand(rbx, rcx, times_4, 10000));
|
|
__ vcmppd(xmm5, xmm4, xmm1, 1);
|
|
__ vcmppd(xmm5, xmm4, Operand(rbx, rcx, times_4, 10000), 1);
|
|
__ vcmpeqpd(xmm5, xmm4, xmm1);
|
|
__ vcmpeqpd(xmm5, xmm4, Operand(rbx, rcx, times_4, 10000));
|
|
__ vcmpltpd(xmm5, xmm4, xmm1);
|
|
__ vcmpltpd(xmm5, xmm4, Operand(rbx, rcx, times_4, 10000));
|
|
__ vcmplepd(xmm5, xmm4, xmm1);
|
|
__ vcmplepd(xmm5, xmm4, Operand(rbx, rcx, times_4, 10000));
|
|
__ vcmpneqpd(xmm5, xmm4, xmm1);
|
|
__ vcmpneqpd(xmm5, xmm4, Operand(rbx, rcx, times_4, 10000));
|
|
__ vcmpnltpd(xmm5, xmm4, xmm1);
|
|
__ vcmpnltpd(xmm5, xmm4, Operand(rbx, rcx, times_4, 10000));
|
|
__ vcmpnlepd(xmm5, xmm4, xmm1);
|
|
__ vcmpnlepd(xmm5, xmm4, Operand(rbx, rcx, times_4, 10000));
|
|
|
|
#define EMIT_SSE_UNOP_AVXINSTR(instruction, notUsed1, notUsed2) \
|
|
__ v##instruction(xmm10, xmm1); \
|
|
__ v##instruction(xmm10, Operand(rbx, rcx, times_4, 10000));
|
|
|
|
SSE_UNOP_INSTRUCTION_LIST(EMIT_SSE_UNOP_AVXINSTR)
|
|
#undef EMIT_SSE_UNOP_AVXINSTR
|
|
|
|
#define EMIT_SSE_BINOP_AVXINSTR(instruction, notUsed1, notUsed2) \
|
|
__ v##instruction(xmm10, xmm5, xmm1); \
|
|
__ v##instruction(xmm10, xmm5, Operand(rbx, rcx, times_4, 10000));
|
|
|
|
SSE_BINOP_INSTRUCTION_LIST(EMIT_SSE_BINOP_AVXINSTR)
|
|
#undef EMIT_SSE_BINOP_AVXINSTR
|
|
|
|
#define EMIT_SSE2_AVXINSTR(instruction, notUsed1, notUsed2, notUsed3) \
|
|
__ v##instruction(xmm10, xmm5, xmm1); \
|
|
__ v##instruction(xmm10, xmm5, Operand(rdx, 4));
|
|
|
|
SSE2_INSTRUCTION_LIST(EMIT_SSE2_AVXINSTR)
|
|
#undef EMIT_SSE2_AVXINSTR
|
|
|
|
#define EMIT_SSE2_UNOP_AVXINSTR(instruction, notUsed1, notUsed2, notUsed3) \
|
|
__ v##instruction(xmm10, xmm1); \
|
|
__ v##instruction(xmm10, Operand(rdx, 4));
|
|
|
|
SSE2_UNOP_INSTRUCTION_LIST(EMIT_SSE2_UNOP_AVXINSTR)
|
|
#undef EMIT_SSE2_AVXINSTR
|
|
|
|
#define EMIT_SSE2_SD_AVXINSTR(instruction, notUsed1, notUsed2, notUsed3) \
|
|
__ v##instruction(xmm10, xmm5, xmm1); \
|
|
__ v##instruction(xmm10, xmm5, Operand(rbx, rcx, times_4, 10000));
|
|
SSE2_INSTRUCTION_LIST_SD(EMIT_SSE2_SD_AVXINSTR)
|
|
#undef EMIT_SSE2_SD_AVXINSTR
|
|
|
|
#define EMIT_SSE34_AVXINSTR(instruction, notUsed1, notUsed2, notUsed3, \
|
|
notUsed4) \
|
|
__ v##instruction(xmm10, xmm5, xmm1); \
|
|
__ v##instruction(xmm10, xmm5, Operand(rdx, 4));
|
|
|
|
SSSE3_INSTRUCTION_LIST(EMIT_SSE34_AVXINSTR)
|
|
SSE4_INSTRUCTION_LIST(EMIT_SSE34_AVXINSTR)
|
|
SSE4_2_INSTRUCTION_LIST(EMIT_SSE34_AVXINSTR)
|
|
#undef EMIT_SSE34_AVXINSTR
|
|
|
|
#define EMIT_SSSE3_UNOP_AVXINSTR(instruction, notUsed1, notUsed2, notUsed3, \
|
|
notUsed4) \
|
|
__ v##instruction(xmm9, xmm3); \
|
|
__ v##instruction(xmm9, Operand(rdx, 5));
|
|
SSSE3_UNOP_INSTRUCTION_LIST(EMIT_SSSE3_UNOP_AVXINSTR)
|
|
#undef EMIT_SSSE3_UNOP_AVXINSTR
|
|
|
|
#define EMIT_SSE4_PMOV_AVXINSTR(instruction, notUsed1, notUsed2, notUsed3, \
|
|
notUsed4) \
|
|
__ v##instruction(xmm10, xmm1); \
|
|
__ v##instruction(xmm10, Operand(rdx, 4));
|
|
SSE4_UNOP_INSTRUCTION_LIST(EMIT_SSE4_PMOV_AVXINSTR)
|
|
#undef EMIT_SSE4_PMOV_AVXINSTR
|
|
|
|
#define EMIT_SSE2_SHIFT_IMM_AVX(instruction, notUsed1, notUsed2, notUsed3, \
|
|
notUsed4) \
|
|
__ v##instruction(xmm0, xmm15, 21);
|
|
SSE2_INSTRUCTION_LIST_SHIFT_IMM(EMIT_SSE2_SHIFT_IMM_AVX)
|
|
#undef EMIT_SSE2_SHIFT_IMM_AVX
|
|
|
|
__ vinsertps(xmm1, xmm2, xmm3, 1);
|
|
__ vinsertps(xmm1, xmm2, Operand(rbx, rcx, times_4, 10000), 1);
|
|
__ vextractps(rax, xmm1, 1);
|
|
|
|
__ vlddqu(xmm1, Operand(rbx, rcx, times_4, 10000));
|
|
__ vpextrb(rax, xmm2, 12);
|
|
__ vpextrb(Operand(rbx, rcx, times_4, 10000), xmm2, 12);
|
|
__ vpextrw(rax, xmm2, 5);
|
|
__ vpextrw(Operand(rbx, rcx, times_4, 10000), xmm2, 5);
|
|
__ vpextrd(rax, xmm2, 2);
|
|
__ vpextrd(Operand(rbx, rcx, times_4, 10000), xmm2, 2);
|
|
__ vpextrq(rax, xmm2, 2);
|
|
|
|
__ vpinsrb(xmm1, xmm2, rax, 12);
|
|
__ vpinsrb(xmm1, xmm2, Operand(rbx, rcx, times_4, 10000), 12);
|
|
__ vpinsrw(xmm1, xmm2, rax, 5);
|
|
__ vpinsrw(xmm1, xmm2, Operand(rbx, rcx, times_4, 10000), 5);
|
|
__ vpinsrd(xmm1, xmm2, rax, 2);
|
|
__ vpinsrd(xmm1, xmm2, Operand(rbx, rcx, times_4, 10000), 2);
|
|
__ vpinsrq(xmm1, xmm2, rax, 9);
|
|
__ vpinsrq(xmm1, xmm2, Operand(rbx, rcx, times_4, 10000), 9);
|
|
__ vpshufd(xmm1, xmm2, 85);
|
|
__ vpshufd(xmm1, Operand(rbx, rcx, times_4, 10000), 85);
|
|
__ vpshuflw(xmm1, xmm2, 85);
|
|
__ vpshuflw(xmm1, Operand(rbx, rcx, times_4, 10000), 85);
|
|
__ vpshufhw(xmm1, xmm2, 85);
|
|
__ vpshufhw(xmm1, Operand(rbx, rcx, times_4, 10000), 85);
|
|
__ vshufps(xmm3, xmm2, xmm3, 3);
|
|
__ vpblendw(xmm1, xmm2, xmm3, 23);
|
|
__ vpblendw(xmm1, xmm2, Operand(rbx, rcx, times_4, 10000), 23);
|
|
__ vpalignr(xmm1, xmm2, xmm3, 4);
|
|
__ vpalignr(xmm1, xmm2, Operand(rbx, rcx, times_4, 10000), 4);
|
|
|
|
__ vpblendvb(xmm1, xmm2, xmm3, xmm4);
|
|
__ vblendvps(xmm1, xmm2, xmm3, xmm4);
|
|
__ vblendvpd(xmm1, xmm2, xmm3, xmm4);
|
|
|
|
__ vmovddup(xmm1, xmm2);
|
|
__ vmovddup(xmm1, Operand(rbx, rcx, times_4, 10000));
|
|
__ vmovshdup(xmm1, xmm2);
|
|
__ vbroadcastss(xmm1, Operand(rbx, rcx, times_4, 10000));
|
|
}
|
|
}
|
|
|
|
// AVX2 instructions.
|
|
{
|
|
if (CpuFeatures::IsSupported(AVX2)) {
|
|
CpuFeatureScope scope(&assm, AVX2);
|
|
#define EMIT_AVX2_BROADCAST(instruction, notUsed1, notUsed2, notUsed3, \
|
|
notUsed4) \
|
|
__ instruction(xmm0, xmm1); \
|
|
__ instruction(xmm0, Operand(rbx, rcx, times_4, 10000));
|
|
AVX2_BROADCAST_LIST(EMIT_AVX2_BROADCAST)
|
|
}
|
|
}
|
|
|
|
// FMA3 instruction
|
|
{
|
|
if (CpuFeatures::IsSupported(FMA3)) {
|
|
CpuFeatureScope scope(&assm, FMA3);
|
|
#define EMIT_FMA(instr, notUsed1, notUsed2, notUsed3, notUsed4, notUsed5, \
|
|
notUsed6) \
|
|
__ instr(xmm9, xmm10, xmm11); \
|
|
__ instr(xmm9, xmm10, Operand(rbx, rcx, times_4, 10000));
|
|
FMA_INSTRUCTION_LIST(EMIT_FMA)
|
|
#undef EMIT_FMA
|
|
}
|
|
}
|
|
|
|
// BMI1 instructions
|
|
{
|
|
if (CpuFeatures::IsSupported(BMI1)) {
|
|
CpuFeatureScope scope(&assm, BMI1);
|
|
__ andnq(rax, rbx, rcx);
|
|
__ andnq(rax, rbx, Operand(rbx, rcx, times_4, 10000));
|
|
__ andnl(rax, rbx, rcx);
|
|
__ andnl(rax, rbx, Operand(rbx, rcx, times_4, 10000));
|
|
__ bextrq(rax, rbx, rcx);
|
|
__ bextrq(rax, Operand(rbx, rcx, times_4, 10000), rbx);
|
|
__ bextrl(rax, rbx, rcx);
|
|
__ bextrl(rax, Operand(rbx, rcx, times_4, 10000), rbx);
|
|
__ blsiq(rax, rbx);
|
|
__ blsiq(rax, Operand(rbx, rcx, times_4, 10000));
|
|
__ blsil(rax, rbx);
|
|
__ blsil(rax, Operand(rbx, rcx, times_4, 10000));
|
|
__ blsmskq(rax, rbx);
|
|
__ blsmskq(rax, Operand(rbx, rcx, times_4, 10000));
|
|
__ blsmskl(rax, rbx);
|
|
__ blsmskl(rax, Operand(rbx, rcx, times_4, 10000));
|
|
__ blsrq(rax, rbx);
|
|
__ blsrq(rax, Operand(rbx, rcx, times_4, 10000));
|
|
__ blsrl(rax, rbx);
|
|
__ blsrl(rax, Operand(rbx, rcx, times_4, 10000));
|
|
__ tzcntq(rax, rbx);
|
|
__ tzcntq(rax, Operand(rbx, rcx, times_4, 10000));
|
|
__ tzcntl(rax, rbx);
|
|
__ tzcntl(rax, Operand(rbx, rcx, times_4, 10000));
|
|
}
|
|
}
|
|
|
|
// LZCNT instructions
|
|
{
|
|
if (CpuFeatures::IsSupported(LZCNT)) {
|
|
CpuFeatureScope scope(&assm, LZCNT);
|
|
__ lzcntq(rax, rbx);
|
|
__ lzcntq(rax, Operand(rbx, rcx, times_4, 10000));
|
|
__ lzcntl(rax, rbx);
|
|
__ lzcntl(rax, Operand(rbx, rcx, times_4, 10000));
|
|
}
|
|
}
|
|
|
|
// POPCNT instructions
|
|
{
|
|
if (CpuFeatures::IsSupported(POPCNT)) {
|
|
CpuFeatureScope scope(&assm, POPCNT);
|
|
__ popcntq(rax, rbx);
|
|
__ popcntq(rax, Operand(rbx, rcx, times_4, 10000));
|
|
__ popcntl(rax, rbx);
|
|
__ popcntl(rax, Operand(rbx, rcx, times_4, 10000));
|
|
}
|
|
}
|
|
|
|
// BMI2 instructions
|
|
{
|
|
if (CpuFeatures::IsSupported(BMI2)) {
|
|
CpuFeatureScope scope(&assm, BMI2);
|
|
__ bzhiq(rax, rbx, rcx);
|
|
__ bzhiq(rax, Operand(rbx, rcx, times_4, 10000), rbx);
|
|
__ bzhil(rax, rbx, rcx);
|
|
__ bzhil(rax, Operand(rbx, rcx, times_4, 10000), rbx);
|
|
__ mulxq(rax, rbx, rcx);
|
|
__ mulxq(rax, rbx, Operand(rbx, rcx, times_4, 10000));
|
|
__ mulxl(rax, rbx, rcx);
|
|
__ mulxl(rax, rbx, Operand(rbx, rcx, times_4, 10000));
|
|
__ pdepq(rax, rbx, rcx);
|
|
__ pdepq(rax, rbx, Operand(rbx, rcx, times_4, 10000));
|
|
__ pdepl(rax, rbx, rcx);
|
|
__ pdepl(rax, rbx, Operand(rbx, rcx, times_4, 10000));
|
|
__ pextq(rax, rbx, rcx);
|
|
__ pextq(rax, rbx, Operand(rbx, rcx, times_4, 10000));
|
|
__ pextl(rax, rbx, rcx);
|
|
__ pextl(rax, rbx, Operand(rbx, rcx, times_4, 10000));
|
|
__ sarxq(rax, rbx, rcx);
|
|
__ sarxq(rax, Operand(rbx, rcx, times_4, 10000), rbx);
|
|
__ sarxl(rax, rbx, rcx);
|
|
__ sarxl(rax, Operand(rbx, rcx, times_4, 10000), rbx);
|
|
__ shlxq(rax, rbx, rcx);
|
|
__ shlxq(rax, Operand(rbx, rcx, times_4, 10000), rbx);
|
|
__ shlxl(rax, rbx, rcx);
|
|
__ shlxl(rax, Operand(rbx, rcx, times_4, 10000), rbx);
|
|
__ shrxq(rax, rbx, rcx);
|
|
__ shrxq(rax, Operand(rbx, rcx, times_4, 10000), rbx);
|
|
__ shrxl(rax, rbx, rcx);
|
|
__ shrxl(rax, Operand(rbx, rcx, times_4, 10000), rbx);
|
|
__ rorxq(rax, rbx, 63);
|
|
__ rorxq(rax, Operand(rbx, rcx, times_4, 10000), 63);
|
|
__ rorxl(rax, rbx, 31);
|
|
__ rorxl(rax, Operand(rbx, rcx, times_4, 10000), 31);
|
|
}
|
|
}
|
|
|
|
// xchg.
|
|
{
|
|
__ xchgb(rax, Operand(rax, 8));
|
|
__ xchgw(rax, Operand(rbx, 8));
|
|
__ xchgq(rax, rax);
|
|
__ xchgq(rax, rbx);
|
|
__ xchgq(rbx, rbx);
|
|
__ xchgq(rbx, Operand(rsp, 12));
|
|
}
|
|
|
|
// cmpxchg.
|
|
{
|
|
__ cmpxchgb(Operand(rsp, 12), rax);
|
|
__ cmpxchgw(Operand(rbx, rcx, times_4, 10000), rax);
|
|
__ cmpxchgl(Operand(rbx, rcx, times_4, 10000), rax);
|
|
__ cmpxchgq(Operand(rbx, rcx, times_4, 10000), rax);
|
|
}
|
|
|
|
// xadd.
|
|
{
|
|
__ xaddb(Operand(rsp, 12), rax);
|
|
__ xaddw(Operand(rsp, 12), rax);
|
|
__ xaddl(Operand(rsp, 12), rax);
|
|
__ xaddq(Operand(rsp, 12), rax);
|
|
__ xaddb(Operand(rbx, rcx, times_4, 10000), rax);
|
|
__ xaddw(Operand(rbx, rcx, times_4, 10000), rax);
|
|
__ xaddl(Operand(rbx, rcx, times_4, 10000), rax);
|
|
__ xaddq(Operand(rbx, rcx, times_4, 10000), rax);
|
|
}
|
|
|
|
// lock prefix.
|
|
{
|
|
__ lock();
|
|
__ cmpxchgl(Operand(rsp, 12), rbx);
|
|
|
|
__ lock();
|
|
__ xchgw(rax, Operand(rcx, 8));
|
|
}
|
|
|
|
// Nop instructions
|
|
for (int i = 0; i < 16; i++) {
|
|
__ Nop(i);
|
|
}
|
|
|
|
__ mfence();
|
|
__ lfence();
|
|
__ pause();
|
|
__ ret(0);
|
|
|
|
CodeDesc desc;
|
|
assm.GetCode(isolate, &desc);
|
|
Handle<Code> code =
|
|
Factory::CodeBuilder(isolate, desc, CodeKind::FOR_TESTING).Build();
|
|
USE(code);
|
|
#ifdef OBJECT_PRINT
|
|
StdoutStream os;
|
|
code->Print(os);
|
|
Address begin = code->raw_instruction_start();
|
|
Address end = code->raw_instruction_end();
|
|
disasm::Disassembler::Disassemble(stdout, reinterpret_cast<byte*>(begin),
|
|
reinterpret_cast<byte*>(end));
|
|
#endif
|
|
}
|
|
|
|
#undef __
|
|
|
|
} // namespace internal
|
|
} // namespace v8
|