~mdw
/
catacomb
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
math/mpx-mul4-*.S: Fix up some of the commentary.
[catacomb]
/
symm
/
gcm-x86ish-pclmul.S
diff --git
a/symm/gcm-x86ish-pclmul.S
b/symm/gcm-x86ish-pclmul.S
index
e60b7ca
..
5edf56e
100644
(file)
--- a/
symm/gcm-x86ish-pclmul.S
+++ b/
symm/gcm-x86ish-pclmul.S
@@
-576,7
+576,7
@@
// xmm3 = // v_0 = (v_01; v_00)
movdqa xmm4, xmm0 // u_1 again
#if CPUFAM_X86
// xmm3 = // v_0 = (v_01; v_00)
movdqa xmm4, xmm0 // u_1 again
#if CPUFAM_X86
- movdqa [
esp
+ 0], xmm3
+ movdqa [
SP
+ 0], xmm3
#elif CPUFAM_AMD64
movdqa xmm8, xmm3
# define V0 xmm8
#elif CPUFAM_AMD64
movdqa xmm8, xmm3
# define V0 xmm8
@@
-608,7
+608,7
@@
pclmullqlqdq xmm4, xmm2 // u_11 v_11
pclmulhqhqdq xmm7, xmm2 // u_10 v_10
#if CPUFAM_X86
pclmullqlqdq xmm4, xmm2 // u_11 v_11
pclmulhqhqdq xmm7, xmm2 // u_10 v_10
#if CPUFAM_X86
- movdqa xmm2, [
esp
+ 0]
+ movdqa xmm2, [
SP
+ 0]
# define V0 xmm2
#endif
pxor xmm0, xmm3 // u_10 v_11 + u_11 v_10
# define V0 xmm2
#endif
pxor xmm0, xmm3 // u_10 v_11 + u_11 v_10
@@
-771,8
+771,8
@@
SSEFUNC(gcm_mulk_128b_x86ish_pclmul)
// A is updated with the product A K.
#if CPUFAM_X86
// A is updated with the product A K.
#if CPUFAM_X86
- mov A, [
esp
+ 4]
- mov K, [
esp
+ 8]
+ mov A, [
SP
+ 4]
+ mov K, [
SP
+ 8]
#endif
endprologue
movdqu xmm0, [A]
#endif
endprologue
movdqu xmm0, [A]
@@
-790,8
+790,8
@@
SSEFUNC(gcm_mulk_128l_x86ish_pclmul)
// exit, A is updated with the product A K.
#if CPUFAM_X86
// exit, A is updated with the product A K.
#if CPUFAM_X86
- mov A, [
esp
+ 4]
- mov K, [
esp
+ 8]
+ mov A, [
SP
+ 4]
+ mov K, [
SP
+ 8]
ldgot ecx
#endif
endprologue
ldgot ecx
#endif
endprologue
@@
-811,8
+811,8
@@
SSEFUNC(gcm_mulk_64b_x86ish_pclmul)
// A is updated with the product A K.
#if CPUFAM_X86
// A is updated with the product A K.
#if CPUFAM_X86
- mov A, [
esp
+ 4]
- mov K, [
esp
+ 8]
+ mov A, [
SP
+ 4]
+ mov K, [
SP
+ 8]
#endif
endprologue
movq xmm0, [A]
#endif
endprologue
movq xmm0, [A]
@@
-830,8
+830,8
@@
SSEFUNC(gcm_mulk_64l_x86ish_pclmul)
// exit, A is updated with the product A K.
#if CPUFAM_X86
// exit, A is updated with the product A K.
#if CPUFAM_X86
- mov A, [
esp
+ 4]
- mov K, [
esp
+ 8]
+ mov A, [
SP
+ 4]
+ mov K, [
SP
+ 8]
ldgot ecx
#endif
endprologue
ldgot ecx
#endif
endprologue
@@
-852,8
+852,8
@@
SSEFUNC(gcm_mulk_96b_x86ish_pclmul)
// with the product A K.
#if CPUFAM_X86
// with the product A K.
#if CPUFAM_X86
- mov A, [
esp
+ 4]
- mov K, [
esp
+ 8]
+ mov A, [
SP
+ 4]
+ mov K, [
SP
+ 8]
#endif
endprologue
movq xmm0, [A + 0]
#endif
endprologue
movq xmm0, [A + 0]
@@
-876,8
+876,8
@@
SSEFUNC(gcm_mulk_96l_x86ish_pclmul)
// updated with the product A K.
#if CPUFAM_X86
// updated with the product A K.
#if CPUFAM_X86
- mov A, [
esp
+ 4]
- mov K, [
esp
+ 8]
+ mov A, [
SP
+ 4]
+ mov K, [
SP
+ 8]
ldgot ecx
#endif
endprologue
ldgot ecx
#endif
endprologue
@@
-901,8
+901,8
@@
SSEFUNC(gcm_mulk_192b_x86ish_pclmul)
// A is updated with the product A K.
#if CPUFAM_X86
// A is updated with the product A K.
#if CPUFAM_X86
- mov A, [
esp
+ 4]
- mov K, [
esp
+ 8]
+ mov A, [
SP
+ 4]
+ mov K, [
SP
+ 8]
#endif
#if CPUFAM_AMD64 && ABI_WIN
stalloc 2*16 + 8
#endif
#if CPUFAM_AMD64 && ABI_WIN
stalloc 2*16 + 8
@@
-935,8
+935,8
@@
SSEFUNC(gcm_mulk_192l_x86ish_pclmul)
// exit, A is updated with the product A K.
#if CPUFAM_X86
// exit, A is updated with the product A K.
#if CPUFAM_X86
- mov A, [
esp
+ 4]
- mov K, [
esp
+ 8]
+ mov A, [
SP
+ 4]
+ mov K, [
SP
+ 8]
ldgot ecx
#endif
#if CPUFAM_AMD64 && ABI_WIN
ldgot ecx
#endif
#if CPUFAM_AMD64 && ABI_WIN
@@
-970,12
+970,12
@@
SSEFUNC(gcm_mulk_256b_x86ish_pclmul)
// A is updated with the product A K.
#if CPUFAM_X86
// A is updated with the product A K.
#if CPUFAM_X86
- pushreg
ebp
+ pushreg
BP
setfp
setfp
- mov A, [
esp
+ 8]
- mov K, [
esp
+ 12]
- and esp, ~15
- sub esp, 16
+ mov A, [
SP
+ 8]
+ mov K, [
SP
+ 12]
+ stalloc 16
+ and SP, ~15
#endif
#if CPUFAM_AMD64 && ABI_WIN
stalloc 3*16 + 8
#endif
#if CPUFAM_AMD64 && ABI_WIN
stalloc 3*16 + 8
@@
-997,7
+997,7
@@
SSEFUNC(gcm_mulk_256b_x86ish_pclmul)
movdqu [A + 0], xmm1
#if CPUFAM_X86
dropfp
movdqu [A + 0], xmm1
#if CPUFAM_X86
dropfp
- popreg
ebp
+ popreg
BP
#endif
#if CPUFAM_AMD64 && ABI_WIN
rstrxmm xmm6, 0
#endif
#if CPUFAM_AMD64 && ABI_WIN
rstrxmm xmm6, 0
@@
-1014,13
+1014,13
@@
SSEFUNC(gcm_mulk_256l_x86ish_pclmul)
// exit, A is updated with the product A K.
#if CPUFAM_X86
// exit, A is updated with the product A K.
#if CPUFAM_X86
- pushreg
ebp
+ pushreg
BP
setfp
setfp
- mov A, [
esp
+ 8]
- mov K, [
esp
+ 12]
- and esp, ~15
+ mov A, [
SP
+ 8]
+ mov K, [
SP
+ 12]
+ stalloc 16
ldgot ecx
ldgot ecx
- sub esp, 16
+ and SP, ~15
#endif
#if CPUFAM_AMD64 && ABI_WIN
stalloc 3*16 + 8
#endif
#if CPUFAM_AMD64 && ABI_WIN
stalloc 3*16 + 8
@@
-1044,7
+1044,7
@@
SSEFUNC(gcm_mulk_256l_x86ish_pclmul)
movdqu [A + 0], xmm1
#if CPUFAM_X86
dropfp
movdqu [A + 0], xmm1
#if CPUFAM_X86
dropfp
- popreg
ebp
+ popreg
BP
#endif
#if CPUFAM_AMD64 && ABI_WIN
rstrxmm xmm6, 0
#endif
#if CPUFAM_AMD64 && ABI_WIN
rstrxmm xmm6, 0