X64: Make more computations happen on 32-bit values instead of on smis.

Review URL: http://codereview.chromium.org/2816014

git-svn-id: http://v8.googlecode.com/svn/branches/bleeding_edge@4899 ce2b1a6d-e550-0410-aec6-3dcde31c8c00
This commit is contained in:
lrn@chromium.org 2010-06-18 12:55:04 +00:00
parent 685cae6021
commit d8bc7a9436
2 changed files with 48 additions and 51 deletions

View File

@ -893,19 +893,20 @@ void KeyedStoreIC::GenerateGeneric(MacroAssembler* masm) {
// -- rdx : receiver // -- rdx : receiver
// -- rsp[0] : return address // -- rsp[0] : return address
// ----------------------------------- // -----------------------------------
Label slow, fast, array, extra, check_pixel_array; Label slow, slow_with_tagged_index, fast, array, extra, check_pixel_array;
// Check that the object isn't a smi. // Check that the object isn't a smi.
__ JumpIfSmi(rdx, &slow); __ JumpIfSmi(rdx, &slow_with_tagged_index);
// Get the map from the receiver. // Get the map from the receiver.
__ movq(rbx, FieldOperand(rdx, HeapObject::kMapOffset)); __ movq(rbx, FieldOperand(rdx, HeapObject::kMapOffset));
// Check that the receiver does not require access checks. We need // Check that the receiver does not require access checks. We need
// to do this because this generic stub does not perform map checks. // to do this because this generic stub does not perform map checks.
__ testb(FieldOperand(rbx, Map::kBitFieldOffset), __ testb(FieldOperand(rbx, Map::kBitFieldOffset),
Immediate(1 << Map::kIsAccessCheckNeeded)); Immediate(1 << Map::kIsAccessCheckNeeded));
__ j(not_zero, &slow); __ j(not_zero, &slow_with_tagged_index);
// Check that the key is a smi. // Check that the key is a smi.
__ JumpIfNotSmi(rcx, &slow); __ JumpIfNotSmi(rcx, &slow_with_tagged_index);
__ SmiToInteger32(rcx, rcx);
__ CmpInstanceType(rbx, JS_ARRAY_TYPE); __ CmpInstanceType(rbx, JS_ARRAY_TYPE);
__ j(equal, &array); __ j(equal, &array);
@ -916,27 +917,30 @@ void KeyedStoreIC::GenerateGeneric(MacroAssembler* masm) {
// Object case: Check key against length in the elements array. // Object case: Check key against length in the elements array.
// rax: value // rax: value
// rdx: JSObject // rdx: JSObject
// rcx: index (as a smi) // rcx: index
__ movq(rbx, FieldOperand(rdx, JSObject::kElementsOffset)); __ movq(rbx, FieldOperand(rdx, JSObject::kElementsOffset));
// Check that the object is in fast mode (not dictionary). // Check that the object is in fast mode (not dictionary).
__ CompareRoot(FieldOperand(rbx, HeapObject::kMapOffset), __ CompareRoot(FieldOperand(rbx, HeapObject::kMapOffset),
Heap::kFixedArrayMapRootIndex); Heap::kFixedArrayMapRootIndex);
__ j(not_equal, &check_pixel_array); __ j(not_equal, &check_pixel_array);
__ SmiCompare(rcx, FieldOperand(rbx, FixedArray::kLengthOffset)); __ SmiCompareInteger32(FieldOperand(rbx, FixedArray::kLengthOffset), rcx);
// rax: value // rax: value
// rbx: FixedArray // rbx: FixedArray
// rcx: index (as a smi) // rcx: index
__ j(below, &fast); __ j(above, &fast);
// Slow case: call runtime. // Slow case: call runtime.
__ bind(&slow); __ bind(&slow);
__ Integer32ToSmi(rcx, rcx);
__ bind(&slow_with_tagged_index);
GenerateRuntimeSetProperty(masm); GenerateRuntimeSetProperty(masm);
// Never returns to here.
// Check whether the elements is a pixel array. // Check whether the elements is a pixel array.
// rax: value // rax: value
// rdx: receiver // rdx: receiver
// rbx: receiver's elements array // rbx: receiver's elements array
// rcx: index (as a smi), zero-extended. // rcx: index, zero-extended.
__ bind(&check_pixel_array); __ bind(&check_pixel_array);
__ CompareRoot(FieldOperand(rbx, HeapObject::kMapOffset), __ CompareRoot(FieldOperand(rbx, HeapObject::kMapOffset),
Heap::kPixelArrayMapRootIndex); Heap::kPixelArrayMapRootIndex);
@ -944,21 +948,20 @@ void KeyedStoreIC::GenerateGeneric(MacroAssembler* masm) {
// Check that the value is a smi. If a conversion is needed call into the // Check that the value is a smi. If a conversion is needed call into the
// runtime to convert and clamp. // runtime to convert and clamp.
__ JumpIfNotSmi(rax, &slow); __ JumpIfNotSmi(rax, &slow);
__ SmiToInteger32(rdi, rcx); __ cmpl(rcx, FieldOperand(rbx, PixelArray::kLengthOffset));
__ cmpl(rdi, FieldOperand(rbx, PixelArray::kLengthOffset));
__ j(above_equal, &slow); __ j(above_equal, &slow);
// No more bailouts to slow case on this path, so key not needed. // No more bailouts to slow case on this path, so key not needed.
__ SmiToInteger32(rcx, rax); __ SmiToInteger32(rdi, rax);
{ // Clamp the value to [0..255]. { // Clamp the value to [0..255].
Label done; Label done;
__ testl(rcx, Immediate(0xFFFFFF00)); __ testl(rdi, Immediate(0xFFFFFF00));
__ j(zero, &done); __ j(zero, &done);
__ setcc(negative, rcx); // 1 if negative, 0 if positive. __ setcc(negative, rdi); // 1 if negative, 0 if positive.
__ decb(rcx); // 0 if negative, 255 if positive. __ decb(rdi); // 0 if negative, 255 if positive.
__ bind(&done); __ bind(&done);
} }
__ movq(rbx, FieldOperand(rbx, PixelArray::kExternalPointerOffset)); __ movq(rbx, FieldOperand(rbx, PixelArray::kExternalPointerOffset));
__ movb(Operand(rbx, rdi, times_1, 0), rcx); __ movb(Operand(rbx, rcx, times_1, 0), rdi);
__ ret(0); __ ret(0);
// Extra capacity case: Check if there is extra capacity to // Extra capacity case: Check if there is extra capacity to
@ -968,14 +971,14 @@ void KeyedStoreIC::GenerateGeneric(MacroAssembler* masm) {
// rax: value // rax: value
// rdx: receiver (a JSArray) // rdx: receiver (a JSArray)
// rbx: receiver's elements array (a FixedArray) // rbx: receiver's elements array (a FixedArray)
// rcx: index (as a smi) // rcx: index
// flags: smicompare (rdx.length(), rbx) // flags: smicompare (rdx.length(), rbx)
__ j(not_equal, &slow); // do not leave holes in the array __ j(not_equal, &slow); // do not leave holes in the array
__ SmiCompare(rcx, FieldOperand(rbx, FixedArray::kLengthOffset)); __ SmiCompareInteger32(FieldOperand(rbx, FixedArray::kLengthOffset), rcx);
__ j(above_equal, &slow); __ j(below_equal, &slow);
// Increment index to get new length. // Increment index to get new length.
__ SmiAddConstant(rdi, rcx, Smi::FromInt(1)); __ leal(rdi, Operand(rcx, 1));
__ movq(FieldOperand(rdx, JSArray::kLengthOffset), rdi); __ Integer32ToSmiField(FieldOperand(rdx, JSArray::kLengthOffset), rdi);
__ jmp(&fast); __ jmp(&fast);
// Array case: Get the length and the elements array from the JS // Array case: Get the length and the elements array from the JS
@ -984,7 +987,7 @@ void KeyedStoreIC::GenerateGeneric(MacroAssembler* masm) {
__ bind(&array); __ bind(&array);
// rax: value // rax: value
// rdx: receiver (a JSArray) // rdx: receiver (a JSArray)
// rcx: index (as a smi) // rcx: index
__ movq(rbx, FieldOperand(rdx, JSObject::kElementsOffset)); __ movq(rbx, FieldOperand(rdx, JSObject::kElementsOffset));
__ CompareRoot(FieldOperand(rbx, HeapObject::kMapOffset), __ CompareRoot(FieldOperand(rbx, HeapObject::kMapOffset),
Heap::kFixedArrayMapRootIndex); Heap::kFixedArrayMapRootIndex);
@ -992,28 +995,23 @@ void KeyedStoreIC::GenerateGeneric(MacroAssembler* masm) {
// Check the key against the length in the array, compute the // Check the key against the length in the array, compute the
// address to store into and fall through to fast case. // address to store into and fall through to fast case.
__ SmiCompare(FieldOperand(rdx, JSArray::kLengthOffset), rcx); __ SmiCompareInteger32(FieldOperand(rdx, JSArray::kLengthOffset), rcx);
__ j(below_equal, &extra); __ j(below_equal, &extra);
// Fast case: Do the store. // Fast case: Do the store.
__ bind(&fast); __ bind(&fast);
// rax: value // rax: value
// rbx: receiver's elements array (a FixedArray) // rbx: receiver's elements array (a FixedArray)
// rcx: index (as a smi) // rcx: index
Label non_smi_value; Label non_smi_value;
__ JumpIfNotSmi(rax, &non_smi_value); __ movq(FieldOperand(rbx, rcx, times_pointer_size, FixedArray::kHeaderSize),
SmiIndex index = masm->SmiToIndex(rcx, rcx, kPointerSizeLog2);
__ movq(FieldOperand(rbx, index.reg, index.scale, FixedArray::kHeaderSize),
rax); rax);
__ JumpIfNotSmi(rax, &non_smi_value);
__ ret(0); __ ret(0);
__ bind(&non_smi_value); __ bind(&non_smi_value);
// Slow case that needs to retain rcx for use by RecordWrite. // Slow case that needs to retain rcx for use by RecordWrite.
// Update write barrier for the elements array address. // Update write barrier for the elements array address.
SmiIndex index2 = masm->SmiToIndex(kScratchRegister, rcx, kPointerSizeLog2);
__ movq(FieldOperand(rbx, index2.reg, index2.scale, FixedArray::kHeaderSize),
rax);
__ movq(rdx, rax); __ movq(rdx, rax);
__ SmiToInteger32(rcx, rcx);
__ RecordWriteNonSmi(rbx, 0, rdx, rcx); __ RecordWriteNonSmi(rbx, 0, rdx, rcx);
__ ret(0); __ ret(0);
} }

View File

@ -985,30 +985,30 @@ Object* CallStubCompiler::CompileArrayPushCall(Object* object,
Label call_builtin, exit, with_write_barrier, attempt_to_grow_elements; Label call_builtin, exit, with_write_barrier, attempt_to_grow_elements;
// Get the array's length into rax and calculate new length. // Get the array's length into rax and calculate new length.
__ movq(rax, FieldOperand(rdx, JSArray::kLengthOffset)); __ SmiToInteger32(rax, FieldOperand(rdx, JSArray::kLengthOffset));
STATIC_ASSERT(FixedArray::kMaxLength < Smi::kMaxValue); STATIC_ASSERT(FixedArray::kMaxLength < Smi::kMaxValue);
__ SmiAddConstant(rax, rax, Smi::FromInt(argc)); __ addl(rax, Immediate(argc));
// Get the element's length into rcx. // Get the element's length into rcx.
__ movq(rcx, FieldOperand(rbx, FixedArray::kLengthOffset)); __ SmiToInteger32(rcx, FieldOperand(rbx, FixedArray::kLengthOffset));
// Check if we could survive without allocation. // Check if we could survive without allocation.
__ SmiCompare(rax, rcx); __ cmpl(rax, rcx);
__ j(greater, &attempt_to_grow_elements); __ j(greater, &attempt_to_grow_elements);
// Save new length. // Save new length.
__ movq(FieldOperand(rdx, JSArray::kLengthOffset), rax); __ Integer32ToSmiField(FieldOperand(rdx, JSArray::kLengthOffset), rax);
// Push the element. // Push the element.
__ movq(rcx, Operand(rsp, argc * kPointerSize)); __ movq(rcx, Operand(rsp, argc * kPointerSize));
SmiIndex index =
masm()->SmiToIndex(kScratchRegister, rax, times_pointer_size);
__ lea(rdx, FieldOperand(rbx, __ lea(rdx, FieldOperand(rbx,
index.reg, index.scale, rax, times_pointer_size,
FixedArray::kHeaderSize - argc * kPointerSize)); FixedArray::kHeaderSize - argc * kPointerSize));
__ movq(Operand(rdx, 0), rcx); __ movq(Operand(rdx, 0), rcx);
// Check if value is a smi. // Check if value is a smi.
__ Integer32ToSmi(rax, rax); // Return new length as smi.
__ JumpIfNotSmi(rcx, &with_write_barrier); __ JumpIfNotSmi(rcx, &with_write_barrier);
__ bind(&exit); __ bind(&exit);
@ -1020,6 +1020,7 @@ Object* CallStubCompiler::CompileArrayPushCall(Object* object,
RecordWriteStub stub(rbx, rdx, rcx); RecordWriteStub stub(rbx, rdx, rcx);
__ CallStub(&stub); __ CallStub(&stub);
__ ret((argc + 1) * kPointerSize); __ ret((argc + 1) * kPointerSize);
__ bind(&attempt_to_grow_elements); __ bind(&attempt_to_grow_elements);
@ -1034,9 +1035,8 @@ Object* CallStubCompiler::CompileArrayPushCall(Object* object,
__ movq(rcx, Operand(rcx, 0)); __ movq(rcx, Operand(rcx, 0));
// Check if it's the end of elements. // Check if it's the end of elements.
index = masm()->SmiToIndex(kScratchRegister, rax, times_pointer_size);
__ lea(rdx, FieldOperand(rbx, __ lea(rdx, FieldOperand(rbx,
index.reg, index.scale, rax, times_pointer_size,
FixedArray::kHeaderSize - argc * kPointerSize)); FixedArray::kHeaderSize - argc * kPointerSize));
__ cmpq(rdx, rcx); __ cmpq(rdx, rcx);
__ j(not_equal, &call_builtin); __ j(not_equal, &call_builtin);
@ -1064,8 +1064,9 @@ Object* CallStubCompiler::CompileArrayPushCall(Object* object,
// Increment element's and array's sizes. // Increment element's and array's sizes.
__ SmiAddConstant(FieldOperand(rbx, FixedArray::kLengthOffset), __ SmiAddConstant(FieldOperand(rbx, FixedArray::kLengthOffset),
Smi::FromInt(kAllocationDelta)); Smi::FromInt(kAllocationDelta));
// Make new length a smi before returning it.
__ Integer32ToSmi(rax, rax);
__ movq(FieldOperand(rdx, JSArray::kLengthOffset), rax); __ movq(FieldOperand(rdx, JSArray::kLengthOffset), rax);
// Elements are in new space, so write barrier is not required. // Elements are in new space, so write barrier is not required.
__ ret((argc + 1) * kPointerSize); __ ret((argc + 1) * kPointerSize);
@ -1128,28 +1129,26 @@ Object* CallStubCompiler::CompileArrayPopCall(Object* object,
__ j(not_equal, &miss); __ j(not_equal, &miss);
// Get the array's length into rcx and calculate new length. // Get the array's length into rcx and calculate new length.
__ movq(rcx, FieldOperand(rdx, JSArray::kLengthOffset)); __ SmiToInteger32(rcx, FieldOperand(rdx, JSArray::kLengthOffset));
__ SmiSubConstant(rcx, rcx, Smi::FromInt(1)); __ subl(rcx, Immediate(1));
__ SmiTest(rcx);
__ j(negative, &return_undefined); __ j(negative, &return_undefined);
// Get the last element. // Get the last element.
__ Move(r9, Factory::the_hole_value()); __ Move(r9, Factory::the_hole_value());
SmiIndex index =
masm()->SmiToIndex(r8, rcx, times_pointer_size);
__ movq(rax, FieldOperand(rbx, __ movq(rax, FieldOperand(rbx,
index.reg, index.scale, rcx, times_pointer_size,
FixedArray::kHeaderSize)); FixedArray::kHeaderSize));
// Check if element is already the hole. // Check if element is already the hole.
__ cmpq(rax, r9); __ cmpq(rax, r9);
// If so, call slow-case to also check prototypes for value.
__ j(equal, &call_builtin); __ j(equal, &call_builtin);
// Set the array's length. // Set the array's length.
__ movq(FieldOperand(rdx, JSArray::kLengthOffset), rcx); __ Integer32ToSmiField(FieldOperand(rdx, JSArray::kLengthOffset), rcx);
// Fill with the hole and return original value.. // Fill with the hole and return original value.
__ movq(FieldOperand(rbx, __ movq(FieldOperand(rbx,
index.reg, index.scale, rcx, times_pointer_size,
FixedArray::kHeaderSize), FixedArray::kHeaderSize),
r9); r9);
__ ret((argc + 1) * kPointerSize); __ ret((argc + 1) * kPointerSize);