mirror of
https://sourceware.org/git/glibc.git
synced 2024-11-14 01:00:07 +00:00
101 lines
2.1 KiB
ArmAsm
101 lines
2.1 KiB
ArmAsm
/* x86-64 __mpn_add_n -- Add two limb vectors of the same length > 0 and store
|
|
sum in a third limb vector.
|
|
Copyright (C) 2006-2015 Free Software Foundation, Inc.
|
|
This file is part of the GNU MP Library.
|
|
|
|
The GNU MP Library is free software; you can redistribute it and/or modify
|
|
it under the terms of the GNU Lesser General Public License as published by
|
|
the Free Software Foundation; either version 2.1 of the License, or (at your
|
|
option) any later version.
|
|
|
|
The GNU MP Library is distributed in the hope that it will be useful, but
|
|
WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY
|
|
or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public
|
|
License for more details.
|
|
|
|
You should have received a copy of the GNU Lesser General Public License
|
|
along with the GNU MP Library; see the file COPYING.LIB. If not,
|
|
see <http://www.gnu.org/licenses/>. */
|
|
|
|
#include "sysdep.h"
|
|
#include "asm-syntax.h"
|
|
|
|
#define rp %rdi
|
|
#define up %rsi
|
|
#define vp %rdx
|
|
#define n %rcx
|
|
#define cy %r8
|
|
|
|
#ifndef func
|
|
# define func __mpn_add_n
|
|
# define ADCSBB adc
|
|
#endif
|
|
|
|
.text
|
|
ENTRY (func)
|
|
xor %r8, %r8
|
|
mov (up), %r10
|
|
mov (vp), %r11
|
|
|
|
lea -8(up,n,8), up
|
|
lea -8(vp,n,8), vp
|
|
lea -16(rp,n,8), rp
|
|
mov %ecx, %eax
|
|
neg n
|
|
and $3, %eax
|
|
je L(b00)
|
|
add %rax, n /* clear low rcx bits for jrcxz */
|
|
cmp $2, %eax
|
|
jl L(b01)
|
|
je L(b10)
|
|
|
|
L(b11): shr %r8 /* set cy */
|
|
jmp L(e11)
|
|
|
|
L(b00): shr %r8 /* set cy */
|
|
mov %r10, %r8
|
|
mov %r11, %r9
|
|
lea 4(n), n
|
|
jmp L(e00)
|
|
|
|
L(b01): shr %r8 /* set cy */
|
|
jmp L(e01)
|
|
|
|
L(b10): shr %r8 /* set cy */
|
|
mov %r10, %r8
|
|
mov %r11, %r9
|
|
jmp L(e10)
|
|
|
|
L(end): ADCSBB %r11, %r10
|
|
mov %r10, 8(rp)
|
|
mov %ecx, %eax /* clear eax, ecx contains 0 */
|
|
adc %eax, %eax
|
|
ret
|
|
|
|
.p2align 4
|
|
L(top):
|
|
mov -24(up,n,8), %r8
|
|
mov -24(vp,n,8), %r9
|
|
ADCSBB %r11, %r10
|
|
mov %r10, -24(rp,n,8)
|
|
L(e00):
|
|
mov -16(up,n,8), %r10
|
|
mov -16(vp,n,8), %r11
|
|
ADCSBB %r9, %r8
|
|
mov %r8, -16(rp,n,8)
|
|
L(e11):
|
|
mov -8(up,n,8), %r8
|
|
mov -8(vp,n,8), %r9
|
|
ADCSBB %r11, %r10
|
|
mov %r10, -8(rp,n,8)
|
|
L(e10):
|
|
mov (up,n,8), %r10
|
|
mov (vp,n,8), %r11
|
|
ADCSBB %r9, %r8
|
|
mov %r8, (rp,n,8)
|
|
L(e01):
|
|
jrcxz L(end)
|
|
lea 4(n), n
|
|
jmp L(top)
|
|
END (func)
|