mirror of
https://sourceware.org/git/glibc.git
synced 2024-12-26 20:51:11 +00:00
115 lines
2.3 KiB
ArmAsm
115 lines
2.3 KiB
ArmAsm
/* x86-64 __mpn_addmul_1 -- Multiply a limb vector with a limb and add
|
|
the result to a second limb vector.
|
|
Copyright (C) 2003,2004,2005,2007,2008,2009 Free Software Foundation, Inc.
|
|
This file is part of the GNU MP Library.
|
|
|
|
The GNU MP Library is free software; you can redistribute it and/or modify
|
|
it under the terms of the GNU Lesser General Public License as published by
|
|
the Free Software Foundation; either version 2.1 of the License, or (at your
|
|
option) any later version.
|
|
|
|
The GNU MP Library is distributed in the hope that it will be useful, but
|
|
WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY
|
|
or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public
|
|
License for more details.
|
|
|
|
You should have received a copy of the GNU Lesser General Public License
|
|
along with the GNU MP Library; see the file COPYING.LIB. If not,
|
|
see <http://www.gnu.org/licenses/>. */
|
|
|
|
#include "sysdep.h"
|
|
#include "asm-syntax.h"
|
|
|
|
#define rp %rdi
|
|
#define up %rsi
|
|
#define n %rdx
|
|
#define v0 %rcx
|
|
|
|
#ifndef func
|
|
# define func __mpn_addmul_1
|
|
# define ADDSUB add
|
|
#endif
|
|
|
|
.text
|
|
ENTRY (func)
|
|
push %rbx
|
|
push %rbp
|
|
lea (%rdx), %rbx
|
|
neg %rbx
|
|
|
|
mov (up), %rax
|
|
mov (rp), %r10
|
|
|
|
lea -16(rp,%rdx,8), rp
|
|
lea (up,%rdx,8), up
|
|
mul %rcx
|
|
|
|
bt $0, %ebx
|
|
jc L(odd)
|
|
|
|
lea (%rax), %r11
|
|
mov 8(up,%rbx,8), %rax
|
|
lea (%rdx), %rbp
|
|
mul %rcx
|
|
add $2, %rbx
|
|
jns L(n2)
|
|
|
|
lea (%rax), %r8
|
|
mov (up,%rbx,8), %rax
|
|
lea (%rdx), %r9
|
|
jmp L(mid)
|
|
|
|
L(odd): add $1, %rbx
|
|
jns L(n1)
|
|
|
|
lea (%rax), %r8
|
|
mov (up,%rbx,8), %rax
|
|
lea (%rdx), %r9
|
|
mul %rcx
|
|
lea (%rax), %r11
|
|
mov 8(up,%rbx,8), %rax
|
|
lea (%rdx), %rbp
|
|
jmp L(e)
|
|
|
|
.p2align 4
|
|
L(top): mul %rcx
|
|
ADDSUB %r8, %r10
|
|
lea (%rax), %r8
|
|
mov (up,%rbx,8), %rax
|
|
adc %r9, %r11
|
|
mov %r10, -8(rp,%rbx,8)
|
|
mov (rp,%rbx,8), %r10
|
|
lea (%rdx), %r9
|
|
adc $0, %rbp
|
|
L(mid): mul %rcx
|
|
ADDSUB %r11, %r10
|
|
lea (%rax), %r11
|
|
mov 8(up,%rbx,8), %rax
|
|
adc %rbp, %r8
|
|
mov %r10, (rp,%rbx,8)
|
|
mov 8(rp,%rbx,8), %r10
|
|
lea (%rdx), %rbp
|
|
adc $0, %r9
|
|
L(e): add $2, %rbx
|
|
js L(top)
|
|
|
|
mul %rcx
|
|
ADDSUB %r8, %r10
|
|
adc %r9, %r11
|
|
mov %r10, -8(rp)
|
|
adc $0, %rbp
|
|
L(n2): mov (rp), %r10
|
|
ADDSUB %r11, %r10
|
|
adc %rbp, %rax
|
|
mov %r10, (rp)
|
|
adc $0, %rdx
|
|
L(n1): mov 8(rp), %r10
|
|
ADDSUB %rax, %r10
|
|
mov %r10, 8(rp)
|
|
mov %ebx, %eax /* zero rax */
|
|
adc %rdx, %rax
|
|
pop %rbp
|
|
pop %rbx
|
|
ret
|
|
END (func)
|