~mdw
/
catacomb
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
math/: Delete some unnecessary blank lines.
[catacomb]
/
math
/
mpx-mul4-amd64-sse2.S
diff --git
a/math/mpx-mul4-amd64-sse2.S
b/math/mpx-mul4-amd64-sse2.S
index
51e94c5
..
bd8ff2f
100644
(file)
--- a/
math/mpx-mul4-amd64-sse2.S
+++ b/
math/mpx-mul4-amd64-sse2.S
@@
-319,7
+319,6
@@
INTFUNC(carryprop)
movdqu [rdi], xmm0
ret
movdqu [rdi], xmm0
ret
-
ENDFUNC
INTFUNC(dmul4)
ENDFUNC
INTFUNC(dmul4)
@@
-357,7
+356,6
@@
INTFUNC(dmul4)
movdqu [rdi], xmm6
ret
movdqu [rdi], xmm6
ret
-
ENDFUNC
INTFUNC(dmla4)
ENDFUNC
INTFUNC(dmla4)
@@
-398,7
+396,6
@@
INTFUNC(dmla4)
movdqu [rdi], xmm6
ret
movdqu [rdi], xmm6
ret
-
ENDFUNC
INTFUNC(mul4zc)
ENDFUNC
INTFUNC(mul4zc)
@@
-429,7
+426,6
@@
INTFUNC(mul4zc)
movdqu [rdi], xmm6
ret
movdqu [rdi], xmm6
ret
-
ENDFUNC
INTFUNC(mul4)
ENDFUNC
INTFUNC(mul4)
@@
-462,7
+458,6
@@
INTFUNC(mul4)
movdqu [rdi], xmm6
ret
movdqu [rdi], xmm6
ret
-
ENDFUNC
INTFUNC(mla4zc)
ENDFUNC
INTFUNC(mla4zc)
@@
-498,7
+493,6
@@
INTFUNC(mla4zc)
movdqu [rdi], xmm6
ret
movdqu [rdi], xmm6
ret
-
ENDFUNC
INTFUNC(mla4)
ENDFUNC
INTFUNC(mla4)
@@
-533,7
+527,6
@@
INTFUNC(mla4)
movdqu [rdi], xmm6
ret
movdqu [rdi], xmm6
ret
-
ENDFUNC
INTFUNC(mmul4)
ENDFUNC
INTFUNC(mmul4)
@@
-557,7
+550,6
@@
INTFUNC(mmul4)
mulcore xmm4, 0, xmm8, xmm9, xmm12, xmm13, xmm14, xmm15
propout xmm7, lo, xmm12, xmm13
jmp 5f
mulcore xmm4, 0, xmm8, xmm9, xmm12, xmm13, xmm14, xmm15
propout xmm7, lo, xmm12, xmm13
jmp 5f
-
ENDFUNC
INTFUNC(mmla4)
ENDFUNC
INTFUNC(mmla4)
@@
-744,7
+736,6
@@
INTFUNC(mont4)
// And, with that, we're done.
movdqu [rdi], xmm6
ret
// And, with that, we're done.
movdqu [rdi], xmm6
ret
-
ENDFUNC
///--------------------------------------------------------------------------
ENDFUNC
///--------------------------------------------------------------------------
@@
-783,7
+774,6
@@
FUNC(mpx_umul4_amd64_sse2)
endprologue
mov DV, rdi
endprologue
mov DV, rdi
-
#endif
#if ABI_WIN
#endif
#if ABI_WIN
@@
-812,7
+802,6
@@
FUNC(mpx_umul4_amd64_sse2)
mov rdi, DV
mov BVL, [SP + 224]
mov rdi, DV
mov BVL, [SP + 224]
-
#endif
// Prepare for the first iteration.
#endif
// Prepare for the first iteration.
@@
-878,7
+867,6
@@
FUNC(mpx_umul4_amd64_sse2)
#endif
#if ABI_WIN
#endif
#if ABI_WIN
-
rstrxmm xmm6, 0
rstrxmm xmm7, 16
rstrxmm xmm8, 32
rstrxmm xmm6, 0
rstrxmm xmm7, 16
rstrxmm xmm8, 32
@@
-893,7
+881,6
@@
FUNC(mpx_umul4_amd64_sse2)
stfree 160 + 8
popreg rdi
popreg rbx
stfree 160 + 8
popreg rdi
popreg rbx
-
#endif
ret
#endif
ret
@@
-946,7
+933,6
@@
FUNC(mpxmont_mul4_amd64_sse2)
endprologue
mov DV, rdi
endprologue
mov DV, rdi
-
#endif
#if ABI_WIN
#endif
#if ABI_WIN
@@
-980,7
+966,6
@@
FUNC(mpxmont_mul4_amd64_sse2)
mov rdi, DV
mov N, [SP + 224]
mov MI, [SP + 232]
mov rdi, DV
mov N, [SP + 224]
mov MI, [SP + 232]
-
#endif
// Establish the expanded operands.
#endif
// Establish the expanded operands.
@@
-1062,7
+1047,6
@@
FUNC(mpxmont_mul4_amd64_sse2)
#endif
#if ABI_WIN
#endif
#if ABI_WIN
-
rstrxmm xmm6, 0
rstrxmm xmm7, 16
rstrxmm xmm8, 32
rstrxmm xmm6, 0
rstrxmm xmm7, 16
rstrxmm xmm8, 32
@@
-1078,7
+1062,6
@@
FUNC(mpxmont_mul4_amd64_sse2)
popreg r12
popreg rdi
popreg rbx
popreg r12
popreg rdi
popreg rbx
-
#endif
ret
#endif
ret
@@
-1134,7
+1117,6
@@
FUNC(mpxmont_redc4_amd64_sse2)
// c rcx r9
#if ABI_SYSV
// c rcx r9
#if ABI_SYSV
-
# define DVL rax
# define DVL4 rsi
# define MI r8
# define DVL rax
# define DVL4 rsi
# define MI r8
@@
-1149,11
+1131,9
@@
FUNC(mpxmont_redc4_amd64_sse2)
endprologue
mov DV, rdi
endprologue
mov DV, rdi
-
#endif
#if ABI_WIN
#endif
#if ABI_WIN
-
# define DVL rax
# define DVL4 rdx
# define MI r10
# define DVL rax
# define DVL4 rdx
# define MI r10
@@
-1184,7
+1164,6
@@
FUNC(mpxmont_redc4_amd64_sse2)
mov rdi, DV
mov MI, [SP + 224]
mov rdi, DV
mov MI, [SP + 224]
-
#endif
// Establish the expanded operands and the blocks-of-4 dv limit.
#endif
// Establish the expanded operands and the blocks-of-4 dv limit.
@@
-1267,7
+1246,6
@@
FUNC(mpxmont_redc4_amd64_sse2)
#endif
#if ABI_WIN
#endif
#if ABI_WIN
-
rstrxmm xmm6, 0
rstrxmm xmm7, 16
rstrxmm xmm8, 32
rstrxmm xmm6, 0
rstrxmm xmm7, 16
rstrxmm xmm8, 32
@@
-1283,7
+1261,6
@@
FUNC(mpxmont_redc4_amd64_sse2)
popreg r12
popreg rdi
popreg rbx
popreg r12
popreg rdi
popreg rbx
-
#endif
ret
#endif
ret